[ 488.583666] env[61911]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61911) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.584051] env[61911]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61911) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.584051] env[61911]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61911) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.584420] env[61911]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 488.682047] env[61911]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61911) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 488.691823] env[61911]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61911) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 489.303437] env[61911]: INFO nova.virt.driver [None req-b0562033-5f4b-4346-85d7-fe087f75265a None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 489.374075] env[61911]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 489.374242] env[61911]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 489.374354] env[61911]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61911) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 492.458584] env[61911]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-7784116f-7163-4552-a17c-d14cf647d926 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.474955] env[61911]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61911) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 492.475168] env[61911]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-6498b3cf-5df0-4645-a550-0e1fffe9ce18 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.506107] env[61911]: INFO oslo_vmware.api [-] Successfully established new session; session ID is c3291. [ 492.506270] env[61911]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.132s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.506791] env[61911]: INFO nova.virt.vmwareapi.driver [None req-b0562033-5f4b-4346-85d7-fe087f75265a None None] VMware vCenter version: 7.0.3 [ 492.510137] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2dc8d25-b002-406e-9fbb-8f990cde29d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.527047] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c938e636-9289-4637-bc02-ecf7d531db0b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.532783] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f50a7e-4de8-4f33-84cb-41622c2ea45c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.539168] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02c66ac-394c-4f22-9b61-c4991571ac9a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.551849] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77bc43b-687c-47e5-a5b3-9946d5328705 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.557533] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62090fe4-151f-490a-ab27-2bd26236f847 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.588853] env[61911]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-ad50c54f-7fe5-4f64-8421-6ba4a47e8478 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.593860] env[61911]: DEBUG nova.virt.vmwareapi.driver [None req-b0562033-5f4b-4346-85d7-fe087f75265a None None] Extension org.openstack.compute already exists. {{(pid=61911) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 492.596440] env[61911]: INFO nova.compute.provider_config [None req-b0562033-5f4b-4346-85d7-fe087f75265a None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 493.100025] env[61911]: DEBUG nova.context [None req-b0562033-5f4b-4346-85d7-fe087f75265a None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),10721fd0-bbdd-49ce-b626-776ded24f73c(cell1) {{(pid=61911) load_cells /opt/stack/nova/nova/context.py:464}} [ 493.102192] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.102416] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.103250] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 493.103751] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Acquiring lock "10721fd0-bbdd-49ce-b626-776ded24f73c" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.103954] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Lock "10721fd0-bbdd-49ce-b626-776ded24f73c" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.104983] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Lock "10721fd0-bbdd-49ce-b626-776ded24f73c" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 493.125809] env[61911]: INFO dbcounter [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Registered counter for database nova_cell0 [ 493.133870] env[61911]: INFO dbcounter [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Registered counter for database nova_cell1 [ 493.136992] env[61911]: DEBUG oslo_db.sqlalchemy.engines [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61911) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 493.137335] env[61911]: DEBUG oslo_db.sqlalchemy.engines [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61911) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 493.142748] env[61911]: ERROR nova.db.main.api [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 493.142748] env[61911]: result = function(*args, **kwargs) [ 493.142748] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 493.142748] env[61911]: return func(*args, **kwargs) [ 493.142748] env[61911]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 493.142748] env[61911]: result = fn(*args, **kwargs) [ 493.142748] env[61911]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 493.142748] env[61911]: return f(*args, **kwargs) [ 493.142748] env[61911]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 493.142748] env[61911]: return db.service_get_minimum_version(context, binaries) [ 493.142748] env[61911]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 493.142748] env[61911]: _check_db_access() [ 493.142748] env[61911]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 493.142748] env[61911]: stacktrace = ''.join(traceback.format_stack()) [ 493.142748] env[61911]: [ 493.143553] env[61911]: ERROR nova.db.main.api [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 493.143553] env[61911]: result = function(*args, **kwargs) [ 493.143553] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 493.143553] env[61911]: return func(*args, **kwargs) [ 493.143553] env[61911]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 493.143553] env[61911]: result = fn(*args, **kwargs) [ 493.143553] env[61911]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 493.143553] env[61911]: return f(*args, **kwargs) [ 493.143553] env[61911]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 493.143553] env[61911]: return db.service_get_minimum_version(context, binaries) [ 493.143553] env[61911]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 493.143553] env[61911]: _check_db_access() [ 493.143553] env[61911]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 493.143553] env[61911]: stacktrace = ''.join(traceback.format_stack()) [ 493.143553] env[61911]: [ 493.144132] env[61911]: WARNING nova.objects.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 493.144132] env[61911]: WARNING nova.objects.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Failed to get minimum service version for cell 10721fd0-bbdd-49ce-b626-776ded24f73c [ 493.144501] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Acquiring lock "singleton_lock" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 493.144669] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Acquired lock "singleton_lock" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 493.144914] env[61911]: DEBUG oslo_concurrency.lockutils [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Releasing lock "singleton_lock" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 493.145243] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Full set of CONF: {{(pid=61911) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 493.145390] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ******************************************************************************** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 493.145520] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Configuration options gathered from: {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 493.145659] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 493.145852] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 493.145981] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ================================================================================ {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 493.146208] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] allow_resize_to_same_host = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.146379] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] arq_binding_timeout = 300 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.146512] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] backdoor_port = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.146642] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] backdoor_socket = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.146843] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] block_device_allocate_retries = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.147019] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] block_device_allocate_retries_interval = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.147196] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cert = self.pem {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.147361] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.147611] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute_monitors = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.147807] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] config_dir = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.147983] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] config_drive_format = iso9660 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.148135] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.148306] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] config_source = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.148477] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] console_host = devstack {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.148642] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] control_exchange = nova {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.148803] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cpu_allocation_ratio = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.148965] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] daemon = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.149147] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] debug = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.149307] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] default_access_ip_network_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.149474] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] default_availability_zone = nova {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.149630] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] default_ephemeral_format = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.149832] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] default_green_pool_size = 1000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.150098] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.150270] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] default_schedule_zone = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.150431] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] disk_allocation_ratio = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.150596] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] enable_new_services = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.150790] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] enabled_apis = ['osapi_compute'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.150959] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] enabled_ssl_apis = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.151143] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] flat_injected = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.151306] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] force_config_drive = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.151466] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] force_raw_images = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.151635] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] graceful_shutdown_timeout = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.151799] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] heal_instance_info_cache_interval = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.152015] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] host = cpu-1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.152194] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.152360] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.152520] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.152752] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.152938] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instance_build_timeout = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.153112] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instance_delete_interval = 300 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.153282] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instance_format = [instance: %(uuid)s] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.153448] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instance_name_template = instance-%08x {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.153611] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instance_usage_audit = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.153782] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instance_usage_audit_period = month {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.153945] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.154123] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.154291] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] internal_service_availability_zone = internal {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.154448] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] key = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.154608] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] live_migration_retry_count = 30 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.154778] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_color = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.154941] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_config_append = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.155120] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.155282] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_dir = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.155438] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.155566] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_options = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.155729] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_rotate_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.155923] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_rotate_interval_type = days {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.156106] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] log_rotation_type = none {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.156240] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.156366] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.156533] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.156697] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.156827] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.156991] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] long_rpc_timeout = 1800 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.157170] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] max_concurrent_builds = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.157332] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] max_concurrent_live_migrations = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.157490] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] max_concurrent_snapshots = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.157756] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] max_local_block_devices = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.157936] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] max_logfile_count = 30 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.158111] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] max_logfile_size_mb = 200 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.158273] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] maximum_instance_delete_attempts = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.158442] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] metadata_listen = 0.0.0.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.158612] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] metadata_listen_port = 8775 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.158826] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] metadata_workers = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.159024] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] migrate_max_retries = -1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.159196] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] mkisofs_cmd = genisoimage {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.159405] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.159538] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] my_ip = 10.180.1.21 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.159741] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.159908] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] network_allocate_retries = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.160097] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.160270] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.160432] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] osapi_compute_listen_port = 8774 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.160602] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] osapi_compute_unique_server_name_scope = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.160920] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] osapi_compute_workers = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.161112] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] password_length = 12 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.161280] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] periodic_enable = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.161442] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] periodic_fuzzy_delay = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.161631] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] pointer_model = usbtablet {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.161832] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] preallocate_images = none {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.161997] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] publish_errors = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.162141] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] pybasedir = /opt/stack/nova {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.162298] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ram_allocation_ratio = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.162457] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] rate_limit_burst = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.162623] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] rate_limit_except_level = CRITICAL {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.162783] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] rate_limit_interval = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.162945] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] reboot_timeout = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.163116] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] reclaim_instance_interval = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.163274] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] record = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.163441] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] reimage_timeout_per_gb = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.163608] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] report_interval = 120 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.163772] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] rescue_timeout = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.163932] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] reserved_host_cpus = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.164101] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] reserved_host_disk_mb = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.164264] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] reserved_host_memory_mb = 512 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.164423] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] reserved_huge_pages = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.164582] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] resize_confirm_window = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.164773] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] resize_fs_using_block_device = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.164943] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] resume_guests_state_on_host_boot = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.165127] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.165293] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] rpc_response_timeout = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.165454] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] run_external_periodic_tasks = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.165622] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] running_deleted_instance_action = reap {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.165786] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.165945] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] running_deleted_instance_timeout = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.166115] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler_instance_sync_interval = 120 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.166284] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_down_time = 720 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.166452] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] servicegroup_driver = db {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.166607] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] shell_completion = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.166767] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] shelved_offload_time = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.166926] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] shelved_poll_interval = 3600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.167500] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] shutdown_timeout = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.167500] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] source_is_ipv6 = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.167500] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ssl_only = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.167690] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.167824] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] sync_power_state_interval = 600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.168084] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] sync_power_state_pool_size = 1000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.168263] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] syslog_log_facility = LOG_USER {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.168425] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] tempdir = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.168586] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] timeout_nbd = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.168754] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] transport_url = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.168916] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] update_resources_interval = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.169087] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] use_cow_images = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.169251] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] use_eventlog = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.169410] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] use_journal = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.169570] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] use_json = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.169727] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] use_rootwrap_daemon = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.169885] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] use_stderr = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.170054] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] use_syslog = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.170214] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vcpu_pin_set = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.170382] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plugging_is_fatal = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.170546] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plugging_timeout = 300 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.170766] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] virt_mkfs = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.170943] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] volume_usage_poll_interval = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.171120] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] watch_log_file = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.171292] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] web = /usr/share/spice-html5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.171478] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.171649] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.171841] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172027] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_concurrency.disable_process_locking = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172311] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172493] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172662] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172834] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173020] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173184] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173368] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.auth_strategy = keystone {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173536] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.compute_link_prefix = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173747] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173947] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.dhcp_domain = novalocal {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174133] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.enable_instance_password = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174303] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.glance_link_prefix = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174470] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174642] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174809] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.instance_list_per_project_cells = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174972] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.list_records_by_skipping_down_cells = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175148] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.local_metadata_per_cell = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175320] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.max_limit = 1000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175491] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.metadata_cache_expiration = 15 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175665] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.neutron_default_tenant_id = default {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175869] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.response_validation = warn {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176070] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.use_neutron_default_nets = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176249] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176417] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176586] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176788] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176974] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.vendordata_dynamic_targets = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177152] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.vendordata_jsonfile_path = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177335] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177528] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.backend = dogpile.cache.memcached {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177697] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.backend_argument = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177872] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.config_prefix = cache.oslo {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178132] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.dead_timeout = 60.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178316] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.debug_cache_backend = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178485] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.enable_retry_client = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178660] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.enable_socket_keepalive = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178830] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.enabled = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178998] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.enforce_fips_mode = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179179] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.expiration_time = 600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179346] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.hashclient_retry_attempts = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179511] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179676] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_dead_retry = 300 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179866] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_password = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180045] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180215] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180380] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_pool_maxsize = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180542] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180728] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_sasl_enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180914] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181094] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181259] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.memcache_username = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181425] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.proxies = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181589] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.redis_db = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181751] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.redis_password = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181921] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182105] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182280] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.redis_server = localhost:6379 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182445] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.redis_socket_timeout = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182604] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.redis_username = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182794] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.retry_attempts = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182977] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.retry_delay = 0.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183158] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.socket_keepalive_count = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183321] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.socket_keepalive_idle = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183484] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.socket_keepalive_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183644] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.tls_allowed_ciphers = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183804] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.tls_cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183961] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.tls_certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184135] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.tls_enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184295] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cache.tls_keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184464] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184638] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.auth_type = password {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184801] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184977] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185152] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185316] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185478] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.cross_az_attach = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185640] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.debug = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185843] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.endpoint_template = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186062] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.http_retries = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186238] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186400] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186575] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.os_region_name = RegionOne {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186742] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186908] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cinder.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187098] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187261] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.cpu_dedicated_set = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187419] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.cpu_shared_set = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187582] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.image_type_exclude_list = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187744] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187914] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188161] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188349] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188524] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188697] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.resource_provider_association_refresh = 300 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188890] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189096] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.shutdown_retry_interval = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189257] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189435] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] conductor.workers = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189614] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] console.allowed_origins = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189777] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] console.ssl_ciphers = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189949] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] console.ssl_minimum_version = default {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190135] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] consoleauth.enforce_session_timeout = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190306] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] consoleauth.token_ttl = 600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190477] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190667] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190832] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190995] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191170] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191333] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191494] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191670] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191848] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192014] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192181] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.region_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192341] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192499] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192667] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.service_type = accelerator {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192831] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193000] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193166] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193327] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193509] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193672] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] cyborg.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193854] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.backend = sqlalchemy {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194033] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.connection = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194204] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.connection_debug = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194372] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.connection_parameters = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194538] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.connection_recycle_time = 3600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194728] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.connection_trace = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194910] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.db_inc_retry_interval = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195093] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.db_max_retries = 20 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195252] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.db_max_retry_interval = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195415] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.db_retry_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195577] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.max_overflow = 50 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195739] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.max_pool_size = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195901] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.max_retries = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196082] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196245] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.mysql_wsrep_sync_wait = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196405] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.pool_timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196566] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.retry_interval = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196724] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.slave_connection = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196885] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.sqlite_synchronous = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197055] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] database.use_db_reconnect = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197237] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.backend = sqlalchemy {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197408] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.connection = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197573] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.connection_debug = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197776] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.connection_parameters = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197959] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.connection_recycle_time = 3600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198201] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.connection_trace = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198393] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.db_inc_retry_interval = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198563] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.db_max_retries = 20 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198729] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.db_max_retry_interval = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198895] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.db_retry_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199068] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.max_overflow = 50 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199234] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.max_pool_size = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199397] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.max_retries = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199568] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199729] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199953] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.pool_timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200151] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.retry_interval = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200316] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.slave_connection = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200482] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] api_database.sqlite_synchronous = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200692] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] devices.enabled_mdev_types = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200880] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201066] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201238] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ephemeral_storage_encryption.enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201407] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201579] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.api_servers = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201744] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201905] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202081] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202244] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202402] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202564] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.debug = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202728] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.default_trusted_certificate_ids = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202919] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.enable_certificate_validation = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203109] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.enable_rbd_download = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203277] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203444] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203606] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203768] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203924] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204099] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.num_retries = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204272] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.rbd_ceph_conf = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204435] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.rbd_connect_timeout = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204604] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.rbd_pool = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204770] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.rbd_user = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204928] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.region_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205097] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205258] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205426] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.service_type = image {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205596] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205757] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205948] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206125] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206307] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206480] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.verify_glance_signatures = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206639] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] glance.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206849] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] guestfs.debug = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207040] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207212] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.auth_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207372] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207533] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207695] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207859] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208020] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208185] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208433] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208616] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208770] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208964] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209155] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.region_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209319] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209479] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209649] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.service_type = shared-file-system {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209817] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.share_apply_policy_timeout = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209980] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210150] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210309] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210468] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210667] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210841] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] manila.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211021] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] mks.enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211377] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211568] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] image_cache.manager_interval = 2400 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211769] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] image_cache.precache_concurrency = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211948] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] image_cache.remove_unused_base_images = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212133] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212305] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212483] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] image_cache.subdirectory_name = _base {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212696] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.api_max_retries = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212832] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.api_retry_interval = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212993] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213171] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.auth_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213332] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213490] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213651] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213815] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.conductor_group = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213972] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214143] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214300] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214461] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214623] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214808] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214969] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215150] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.peer_list = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215312] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.region_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215470] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215638] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.serial_console_state_timeout = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215797] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215968] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.service_type = baremetal {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216139] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.shard = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216302] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216463] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216620] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216779] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216957] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217129] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ironic.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217313] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217488] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] key_manager.fixed_key = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217695] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217881] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.barbican_api_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218057] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.barbican_endpoint = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218235] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.barbican_endpoint_type = public {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218476] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.barbican_region_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218662] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218822] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218988] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219164] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219324] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219488] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.number_of_retries = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219651] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.retry_delay = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219819] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.send_service_user_token = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219983] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220155] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220318] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.verify_ssl = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220476] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican.verify_ssl_path = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220663] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220868] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.auth_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221046] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221211] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221378] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221543] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221704] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221868] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222039] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] barbican_service_user.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222210] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.approle_role_id = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222371] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.approle_secret_id = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222541] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.kv_mountpoint = secret {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222703] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.kv_path = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222868] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.kv_version = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223043] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.namespace = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223210] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.root_token_id = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223368] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.ssl_ca_crt_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223535] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.timeout = 60.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223707] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.use_ssl = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223899] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224083] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224250] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.auth_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224411] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224569] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224734] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224893] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225061] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225255] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225383] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225541] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225697] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225854] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226023] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.region_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226177] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226385] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226516] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.service_type = identity {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226718] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226841] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227015] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227182] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227363] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227527] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] keystone.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227717] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.ceph_mount_options = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228374] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228658] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.connection_uri = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228841] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.cpu_mode = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229099] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229290] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.cpu_models = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229469] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.cpu_power_governor_high = performance {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229643] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229849] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.cpu_power_management = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230045] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230218] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.device_detach_attempts = 8 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230386] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.device_detach_timeout = 20 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230562] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.disk_cachemodes = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230761] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.disk_prefix = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230937] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.enabled_perf_events = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231119] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.file_backed_memory = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231290] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.gid_maps = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231452] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.hw_disk_discard = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231626] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.hw_machine_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231830] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.images_rbd_ceph_conf = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232011] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232184] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232355] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.images_rbd_glance_store_name = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232527] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.images_rbd_pool = rbd {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232704] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.images_type = default {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232910] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.images_volume_group = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233097] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.inject_key = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233267] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.inject_partition = -2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233432] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.inject_password = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233597] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.iscsi_iface = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233763] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.iser_use_multipath = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233932] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234112] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234280] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_downtime = 500 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234445] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234611] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234782] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_inbound_addr = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234950] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235128] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235293] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_scheme = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235465] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_timeout_action = abort {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235628] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_tunnelled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235813] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_uri = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235992] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.live_migration_with_native_tls = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236170] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.max_queues = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236337] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236562] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236728] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.nfs_mount_options = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237016] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237197] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237364] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237528] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237702] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237907] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.num_pcie_ports = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.238093] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.238268] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.pmem_namespaces = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.238432] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.quobyte_client_cfg = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.238810] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.239012] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.239193] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.239364] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.239528] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rbd_secret_uuid = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.239691] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rbd_user = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.239860] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.240051] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.240227] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rescue_image_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.240389] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rescue_kernel_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.240550] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rescue_ramdisk_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.240743] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.240910] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.rx_queue_size = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.241096] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.smbfs_mount_options = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.241427] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.241630] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.snapshot_compression = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.241803] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.snapshot_image_format = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.242036] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.242212] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.sparse_logical_volumes = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.242380] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.swtpm_enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.242553] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.swtpm_group = tss {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.242724] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.swtpm_user = tss {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.242895] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.sysinfo_serial = unique {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.243068] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.tb_cache_size = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.243231] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.tx_queue_size = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.243395] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.uid_maps = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.243561] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.use_virtio_for_bridges = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.243733] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.virt_type = kvm {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.243904] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.volume_clear = zero {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.244080] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.volume_clear_size = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.244251] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.volume_use_multipath = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.244475] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.vzstorage_cache_path = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.244674] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.244848] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.245027] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.245205] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.245480] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.245659] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.vzstorage_mount_user = stack {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.245829] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.246016] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.246191] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.auth_type = password {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.246354] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.246514] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.246677] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.246839] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.246997] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.247182] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.default_floating_pool = public {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.247352] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.247538] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.extension_sync_interval = 600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.247706] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.http_retries = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.247909] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.248039] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.248203] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.248375] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.248534] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.248779] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.ovs_bridge = br-int {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.248966] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.physnets = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.249153] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.region_name = RegionOne {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.249317] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.249486] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.service_metadata_proxy = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.249647] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.249817] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.service_type = network {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.249985] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.250156] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.250314] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.250497] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.250713] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.250880] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] neutron.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.251065] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] notifications.bdms_in_notifications = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.251249] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] notifications.default_level = INFO {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.251426] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] notifications.notification_format = unversioned {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.251618] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] notifications.notify_on_state_change = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.251768] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.251944] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] pci.alias = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.252126] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] pci.device_spec = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.252293] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] pci.report_in_placement = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.252467] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.252640] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.auth_type = password {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.252808] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.252966] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.253137] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.253304] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.253495] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.253661] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.253824] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.default_domain_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.253984] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.default_domain_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.254157] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.domain_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.254316] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.domain_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.254475] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.254637] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.254796] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.254953] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.255121] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.255290] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.password = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.255450] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.project_domain_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.255618] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.project_domain_name = Default {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.255788] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.project_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.255960] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.project_name = service {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.256142] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.region_name = RegionOne {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.256306] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.256502] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.256688] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.service_type = placement {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.256855] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.257029] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.257227] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.257391] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.system_scope = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.257552] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.257712] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.trust_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.257874] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.user_domain_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.258054] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.user_domain_name = Default {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.258216] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.user_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.258392] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.username = nova {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.258575] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.258738] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] placement.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.259014] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.cores = 20 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.259212] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.count_usage_from_placement = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.259411] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.259605] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.injected_file_content_bytes = 10240 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.259781] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.injected_file_path_length = 255 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.259960] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.injected_files = 5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.260137] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.instances = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.260306] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.key_pairs = 100 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.260475] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.metadata_items = 128 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.260660] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.ram = 51200 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.260835] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.recheck_quota = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.261013] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.server_group_members = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.261189] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] quota.server_groups = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.261364] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.261532] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.261721] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.image_metadata_prefilter = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.261896] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.262072] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.max_attempts = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.262241] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.max_placement_results = 1000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.262409] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.262573] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.262736] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.262911] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] scheduler.workers = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.263095] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.263270] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.263451] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.263619] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.263787] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.263950] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.264127] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.264319] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.264489] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.host_subset_size = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.264681] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.264864] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.265042] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.265215] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.isolated_hosts = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.265382] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.isolated_images = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.265549] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.265711] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.265878] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.266052] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.pci_in_placement = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.266219] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.266394] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.266556] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.266718] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.266880] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.267050] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.267217] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.track_instance_changes = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.267394] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.267566] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] metrics.required = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.267776] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] metrics.weight_multiplier = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.267961] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.268145] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] metrics.weight_setting = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.268468] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.268650] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] serial_console.enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.268826] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] serial_console.port_range = 10000:20000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.269097] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.269285] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.269480] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] serial_console.serialproxy_port = 6083 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.269656] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.269844] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.auth_type = password {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.270028] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.270196] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.270376] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.270550] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.270750] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.270934] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.send_service_user_token = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.271114] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.271276] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] service_user.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.271450] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.agent_enabled = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.271629] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.271935] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.272140] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.272314] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.html5proxy_port = 6082 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.272479] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.image_compression = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.272638] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.jpeg_compression = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.272800] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.playback_compression = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.272982] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.require_secure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.273181] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.server_listen = 127.0.0.1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.273354] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.273516] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.streaming_mode = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.273703] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] spice.zlib_compression = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.273878] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] upgrade_levels.baseapi = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.274064] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] upgrade_levels.compute = auto {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.274229] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] upgrade_levels.conductor = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.274391] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] upgrade_levels.scheduler = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.274559] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.274722] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.274886] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.275053] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.275220] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.275382] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.275542] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.275703] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.275862] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vendordata_dynamic_auth.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.276045] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.api_retry_count = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.276211] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.ca_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.276382] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.276549] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.cluster_name = testcl1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.276750] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.connection_pool_size = 10 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.276922] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.console_delay_seconds = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.277144] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.datastore_regex = ^datastore.* {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.277366] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.277548] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.host_password = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.277715] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.host_port = 443 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.277885] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.host_username = administrator@vsphere.local {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.278067] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.insecure = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.278235] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.integration_bridge = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.278396] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.maximum_objects = 100 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.278553] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.pbm_default_policy = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.278715] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.pbm_enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.278876] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.pbm_wsdl_location = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.279135] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.279315] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.serial_port_proxy_uri = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.279479] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.serial_port_service_uri = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.279667] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.task_poll_interval = 0.5 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.279872] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.use_linked_clone = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.280062] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.vnc_keymap = en-us {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.280236] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.vnc_port = 5900 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.280405] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vmware.vnc_port_total = 10000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.280622] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.auth_schemes = ['none'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.280798] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.281120] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.281311] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.281484] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.novncproxy_port = 6080 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.281660] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.server_listen = 127.0.0.1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.281833] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.281997] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.vencrypt_ca_certs = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.282170] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.vencrypt_client_cert = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.282330] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vnc.vencrypt_client_key = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.282505] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.282671] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.disable_deep_image_inspection = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.282837] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.283025] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.283203] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.283366] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.disable_rootwrap = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.283530] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.enable_numa_live_migration = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.283692] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.283855] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.284025] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.284192] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.libvirt_disable_apic = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.284354] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.284518] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.284677] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.284840] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.285009] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.285180] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.285340] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.285499] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.285658] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.285825] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.286019] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.286195] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.client_socket_timeout = 900 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.286391] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.default_pool_size = 1000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.286565] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.keep_alive = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.286734] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.max_header_line = 16384 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.286899] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.287069] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.ssl_ca_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.287234] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.ssl_cert_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.287395] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.ssl_key_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.287559] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.tcp_keepidle = 600 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.287739] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.287909] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] zvm.ca_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.288080] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] zvm.cloud_connector_url = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.288372] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.288550] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] zvm.reachable_timeout = 300 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.288733] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.enforce_new_defaults = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.289219] env[61911]: WARNING oslo_config.cfg [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 493.289428] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.enforce_scope = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.289613] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.policy_default_rule = default {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.289801] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.289980] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.policy_file = policy.yaml {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.290168] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.290334] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.290497] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.290687] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.290854] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.291036] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.291219] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.291399] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.connection_string = messaging:// {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.291569] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.enabled = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.291741] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.es_doc_type = notification {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.291911] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.es_scroll_size = 10000 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.292093] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.es_scroll_time = 2m {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.292258] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.filter_error_trace = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.292428] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.hmac_keys = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.292595] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.sentinel_service_name = mymaster {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.292761] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.socket_timeout = 0.1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.292933] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.trace_requests = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.293126] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler.trace_sqlalchemy = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.293314] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler_jaeger.process_tags = {} {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.293475] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler_jaeger.service_name_prefix = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.293639] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] profiler_otlp.service_name_prefix = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.293808] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] remote_debug.host = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.293966] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] remote_debug.port = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.294157] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.294324] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.294492] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.294657] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.294824] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.294987] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.295161] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.295325] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.295487] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.295657] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.295819] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.295995] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.296169] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.296340] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.296510] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.296677] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.296841] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.297024] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.297193] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.297353] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.297520] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.297684] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.297849] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.298023] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.298193] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.298356] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.298523] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.298681] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.298849] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.299020] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.ssl = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.299270] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.299465] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.299634] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.299811] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.299983] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.300165] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.300358] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.300531] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_notifications.retry = -1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.300741] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.300930] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.301138] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.auth_section = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.301329] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.auth_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.301495] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.cafile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.301655] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.certfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.301820] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.collect_timing = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.301978] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.connect_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.302150] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.connect_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.302308] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.endpoint_id = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.302466] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.endpoint_override = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.302627] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.insecure = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.302785] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.keyfile = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.302942] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.max_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.303109] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.min_version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.303269] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.region_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.303431] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.retriable_status_codes = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.303590] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.service_name = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.303748] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.service_type = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.303910] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.split_loggers = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.304079] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.status_code_retries = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.304243] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.status_code_retry_delay = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.304401] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.timeout = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.304560] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.valid_interfaces = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.304757] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_limit.version = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.304943] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_reports.file_event_handler = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.305126] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.305292] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] oslo_reports.log_dir = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.305464] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.305627] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.305789] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.305954] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.306130] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.306291] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.306461] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.306621] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_ovs_privileged.group = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.306781] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.306947] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.307120] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.307280] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] vif_plug_ovs_privileged.user = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.307451] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.307630] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.307809] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.307984] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.308174] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.308352] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.308566] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.308711] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.308897] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.309084] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_ovs.isolate_vif = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.309326] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.309521] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.309700] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.309874] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.310051] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] os_vif_ovs.per_port_bridge = False {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.310232] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] privsep_osbrick.capabilities = [21] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.310394] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] privsep_osbrick.group = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.310556] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] privsep_osbrick.helper_command = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.310749] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.310923] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.311094] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] privsep_osbrick.user = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.311273] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.311436] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] nova_sys_admin.group = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.311593] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] nova_sys_admin.helper_command = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.311758] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.311924] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.312091] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] nova_sys_admin.user = None {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.312226] env[61911]: DEBUG oslo_service.service [None req-008713be-bf58-4576-8d2b-392a98f3d755 None None] ******************************************************************************** {{(pid=61911) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 493.312710] env[61911]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 493.816768] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Getting list of instances from cluster (obj){ [ 493.816768] env[61911]: value = "domain-c8" [ 493.816768] env[61911]: _type = "ClusterComputeResource" [ 493.816768] env[61911]: } {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 493.818031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365d0971-bb35-4a03-a208-66ec036ab68f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 493.826893] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Got total of 0 instances {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 493.827432] env[61911]: WARNING nova.virt.vmwareapi.driver [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 493.827906] env[61911]: INFO nova.virt.node [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Generated node identity b8a56394-897a-4e67-92ba-db46db6115e9 [ 493.828158] env[61911]: INFO nova.virt.node [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Wrote node identity b8a56394-897a-4e67-92ba-db46db6115e9 to /opt/stack/data/n-cpu-1/compute_id [ 494.330606] env[61911]: WARNING nova.compute.manager [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Compute nodes ['b8a56394-897a-4e67-92ba-db46db6115e9'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 495.336650] env[61911]: INFO nova.compute.manager [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 496.342490] env[61911]: WARNING nova.compute.manager [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 496.342841] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.342975] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.343150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.343304] env[61911]: DEBUG nova.compute.resource_tracker [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 496.344254] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296ee371-23dd-4818-bb03-cbdac9055bfe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.352567] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2f7141-1abf-4926-852d-9f158dc5adfb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.367006] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2037e5-d661-4397-8702-9f0395874e37 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.373543] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4867eb-0fb4-42b2-8fd7-3b8ac1f7ee9b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.402152] env[61911]: DEBUG nova.compute.resource_tracker [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181510MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 496.402347] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.402589] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.905768] env[61911]: WARNING nova.compute.resource_tracker [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] No compute node record for cpu-1:b8a56394-897a-4e67-92ba-db46db6115e9: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host b8a56394-897a-4e67-92ba-db46db6115e9 could not be found. [ 497.410212] env[61911]: INFO nova.compute.resource_tracker [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: b8a56394-897a-4e67-92ba-db46db6115e9 [ 498.919335] env[61911]: DEBUG nova.compute.resource_tracker [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 498.919699] env[61911]: DEBUG nova.compute.resource_tracker [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 499.074190] env[61911]: INFO nova.scheduler.client.report [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] [req-d6a5d1c5-de5e-4350-bf63-77c77fe3ecb7] Created resource provider record via placement API for resource provider with UUID b8a56394-897a-4e67-92ba-db46db6115e9 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 499.090421] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7bb9e8-8d2c-43aa-ad56-fcfee90ea6bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.097915] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044a5abb-f95a-478a-852e-5969a0693e36 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.127746] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b702b0-89eb-4929-bf7d-f2f3fc29dcb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.134428] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b3cdab-07f0-478e-9261-9b2da5a59c36 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.147081] env[61911]: DEBUG nova.compute.provider_tree [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 499.688613] env[61911]: DEBUG nova.scheduler.client.report [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Updated inventory for provider b8a56394-897a-4e67-92ba-db46db6115e9 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 499.688836] env[61911]: DEBUG nova.compute.provider_tree [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Updating resource provider b8a56394-897a-4e67-92ba-db46db6115e9 generation from 0 to 1 during operation: update_inventory {{(pid=61911) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 499.688999] env[61911]: DEBUG nova.compute.provider_tree [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 499.736762] env[61911]: DEBUG nova.compute.provider_tree [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Updating resource provider b8a56394-897a-4e67-92ba-db46db6115e9 generation from 1 to 2 during operation: update_traits {{(pid=61911) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 500.241350] env[61911]: DEBUG nova.compute.resource_tracker [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 500.241734] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.839s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.241734] env[61911]: DEBUG nova.service [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Creating RPC server for service compute {{(pid=61911) start /opt/stack/nova/nova/service.py:186}} [ 500.256877] env[61911]: DEBUG nova.service [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] Join ServiceGroup membership for this service compute {{(pid=61911) start /opt/stack/nova/nova/service.py:203}} [ 500.257129] env[61911]: DEBUG nova.servicegroup.drivers.db [None req-d179851b-9e68-407c-92ec-b46c49f53855 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61911) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 538.414040] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Acquiring lock "a5021d0d-325d-422f-a3db-eb51b129b72b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.414368] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Lock "a5021d0d-325d-422f-a3db-eb51b129b72b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.729534] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Acquiring lock "02ca2bcc-3a06-4e3d-9985-c158b541cdcc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.729635] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Lock "02ca2bcc-3a06-4e3d-9985-c158b541cdcc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.920606] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 539.103183] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "54cd80a2-17fb-446d-b25e-d6d68f8bd538" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.105489] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "54cd80a2-17fb-446d-b25e-d6d68f8bd538" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.205358] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "82e22142-4083-4716-95a9-dea2d0faf998" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.205598] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "82e22142-4083-4716-95a9-dea2d0faf998" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.233344] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 539.455275] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.455275] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.464449] env[61911]: INFO nova.compute.claims [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.607481] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 539.708176] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 539.816546] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.946193] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Acquiring lock "2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.946546] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Lock "2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.135795] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.235596] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Acquiring lock "2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.235596] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Lock "2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.237400] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.310346] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "960d5c25-2a54-4eb8-b175-27e56b9894d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.310610] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "960d5c25-2a54-4eb8-b175-27e56b9894d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.449163] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 540.603515] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b7a575-3a8b-49bd-9ae4-dc18b3197e84 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.611941] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2942ce-8b2e-447b-8890-8ce17f1cbab5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.642678] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ae2f07-26ed-4283-a2c4-8ba1e5cbf499 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.649880] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4f91ab-13d7-4630-afe7-0fd7dcc7905b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.663117] env[61911]: DEBUG nova.compute.provider_tree [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.741196] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 540.813073] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 540.989578] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.073857] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Acquiring lock "2a8c5d4f-6e2e-452e-848c-69e09c991124" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.073857] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Lock "2a8c5d4f-6e2e-452e-848c-69e09c991124" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.165839] env[61911]: DEBUG nova.scheduler.client.report [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 541.262130] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.336622] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.577705] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 541.671806] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.216s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.672321] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 541.674922] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.859s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.677866] env[61911]: INFO nova.compute.claims [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.701442] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "20ca33cb-1bf2-443a-9a35-31dcd01b4266" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.701824] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "20ca33cb-1bf2-443a-9a35-31dcd01b4266" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.106158] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.186694] env[61911]: DEBUG nova.compute.utils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.187112] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 542.187743] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 542.206490] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 542.404474] env[61911]: DEBUG nova.policy [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28d4c0dccdd54ebc9a1834ef5f4f77f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c4e2ed81c804e26a710bd9bb36330a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.694639] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 542.749320] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.877873] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67853978-4e72-4464-a189-be540e518e04 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.885905] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bffbdf-0c16-4150-bd6c-a7b417e32cbf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.925983] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4793cad8-3528-46db-8b5e-25631e4f7282 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.934514] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b5265a-5b3c-48c2-b878-7c921627145a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.950445] env[61911]: DEBUG nova.compute.provider_tree [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.239506] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Successfully created port: 9df9710d-cc80-4012-a4a7-00c351ecd8f2 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 543.456237] env[61911]: DEBUG nova.scheduler.client.report [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 543.709469] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 543.744497] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 543.744806] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.745036] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 543.745269] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.745565] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 543.746284] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 543.746514] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 543.746802] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 543.747303] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 543.747541] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 543.747764] env[61911]: DEBUG nova.virt.hardware [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 543.748922] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de6b08d-1053-4cc4-868c-e87d4ea03c01 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.758620] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f50253b-0d80-4a75-8dca-d1d37c596481 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.779022] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afb3e36-b8b8-4b06-85e1-8ff96d347e3a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.968509] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.969176] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 543.973254] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.838s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.976187] env[61911]: INFO nova.compute.claims [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.481102] env[61911]: DEBUG nova.compute.utils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.484654] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 544.484922] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.622525] env[61911]: DEBUG nova.policy [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e7aa1f5fcaf4040965455ccc8581858', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e3a076807764a10b71d06c380a767c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.989988] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 545.211457] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6def94cb-cf4f-42dc-b9ee-aacd549dee02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.224363] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba8a129-b45c-4a47-a070-bcf2902c8605 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.257096] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e141c4-e396-42e6-a42a-7bd752b09606 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.268580] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22fd800-6e34-4820-a1c2-d5cc9a60f743 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.283644] env[61911]: DEBUG nova.compute.provider_tree [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.563532] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Successfully created port: bddd1214-67ff-49f9-a0bb-e01cb3b8fc05 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.789956] env[61911]: DEBUG nova.scheduler.client.report [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 546.007832] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 546.035672] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 546.035993] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.036070] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 546.036229] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.036630] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 546.036807] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 546.037022] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 546.037803] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 546.037803] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 546.037803] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 546.037803] env[61911]: DEBUG nova.virt.hardware [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 546.038801] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccaa5eb3-5e0e-410b-8d00-eeccff7fd74b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.048735] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b683379-fced-4a2c-93a5-c95a6d5a197f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.133140] env[61911]: ERROR nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. [ 546.133140] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 546.133140] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 546.133140] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 546.133140] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.133140] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 546.133140] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.133140] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 546.133140] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.133140] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 546.133140] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.133140] env[61911]: ERROR nova.compute.manager raise self.value [ 546.133140] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.133140] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 546.133140] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.133140] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 546.134608] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.134608] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 546.134608] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. [ 546.134608] env[61911]: ERROR nova.compute.manager [ 546.134608] env[61911]: Traceback (most recent call last): [ 546.134608] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 546.134608] env[61911]: listener.cb(fileno) [ 546.134608] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.134608] env[61911]: result = function(*args, **kwargs) [ 546.134608] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.134608] env[61911]: return func(*args, **kwargs) [ 546.134608] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 546.134608] env[61911]: raise e [ 546.134608] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 546.134608] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 546.134608] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.134608] env[61911]: created_port_ids = self._update_ports_for_instance( [ 546.134608] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.134608] env[61911]: with excutils.save_and_reraise_exception(): [ 546.134608] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.134608] env[61911]: self.force_reraise() [ 546.134608] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.134608] env[61911]: raise self.value [ 546.134608] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.134608] env[61911]: updated_port = self._update_port( [ 546.134608] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.134608] env[61911]: _ensure_no_port_binding_failure(port) [ 546.134608] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.134608] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 546.135607] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. [ 546.135607] env[61911]: Removing descriptor: 15 [ 546.135607] env[61911]: ERROR nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Traceback (most recent call last): [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] yield resources [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self.driver.spawn(context, instance, image_meta, [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.135607] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] vm_ref = self.build_virtual_machine(instance, [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] for vif in network_info: [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return self._sync_wrapper(fn, *args, **kwargs) [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self.wait() [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self[:] = self._gt.wait() [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return self._exit_event.wait() [ 546.136195] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] result = hub.switch() [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return self.greenlet.switch() [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] result = function(*args, **kwargs) [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return func(*args, **kwargs) [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] raise e [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] nwinfo = self.network_api.allocate_for_instance( [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.136543] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] created_port_ids = self._update_ports_for_instance( [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] with excutils.save_and_reraise_exception(): [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self.force_reraise() [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] raise self.value [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] updated_port = self._update_port( [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] _ensure_no_port_binding_failure(port) [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.136984] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] raise exception.PortBindingFailed(port_id=port['id']) [ 546.137338] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] nova.exception.PortBindingFailed: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. [ 546.137338] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] [ 546.137338] env[61911]: INFO nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Terminating instance [ 546.295857] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.296494] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 546.301971] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.064s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.303878] env[61911]: INFO nova.compute.claims [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.646140] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Acquiring lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.650016] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Acquired lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.650016] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.815807] env[61911]: DEBUG nova.compute.utils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.823382] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 546.823524] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.955687] env[61911]: DEBUG nova.policy [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36d46402111342708adbc92db5637d2a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88869d5f848f42ccaadc9a80740cf2ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.193095] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.261037] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.305219] env[61911]: DEBUG nova.compute.manager [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Received event network-changed-9df9710d-cc80-4012-a4a7-00c351ecd8f2 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 547.305219] env[61911]: DEBUG nova.compute.manager [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Refreshing instance network info cache due to event network-changed-9df9710d-cc80-4012-a4a7-00c351ecd8f2. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 547.305219] env[61911]: DEBUG oslo_concurrency.lockutils [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] Acquiring lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.324415] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 547.586025] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01aefcd-93d6-4acc-9f3f-06233fa31ad1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.594575] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f5323b-4f15-4371-a65f-8d0b0029e42c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.628805] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51f504a-ac80-4eeb-8b38-97a6068dd058 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.636737] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb241d03-8e66-48c2-987f-5a34bf62fe13 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.652046] env[61911]: DEBUG nova.compute.provider_tree [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.684317] env[61911]: ERROR nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. [ 547.684317] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.684317] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.684317] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.684317] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.684317] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.684317] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.684317] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.684317] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.684317] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 547.684317] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.684317] env[61911]: ERROR nova.compute.manager raise self.value [ 547.684317] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.684317] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.684317] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.684317] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.684989] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.684989] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.684989] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. [ 547.684989] env[61911]: ERROR nova.compute.manager [ 547.684989] env[61911]: Traceback (most recent call last): [ 547.684989] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.684989] env[61911]: listener.cb(fileno) [ 547.684989] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.684989] env[61911]: result = function(*args, **kwargs) [ 547.684989] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.684989] env[61911]: return func(*args, **kwargs) [ 547.684989] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.684989] env[61911]: raise e [ 547.684989] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.684989] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 547.684989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.684989] env[61911]: created_port_ids = self._update_ports_for_instance( [ 547.684989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.684989] env[61911]: with excutils.save_and_reraise_exception(): [ 547.684989] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.684989] env[61911]: self.force_reraise() [ 547.684989] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.684989] env[61911]: raise self.value [ 547.684989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.684989] env[61911]: updated_port = self._update_port( [ 547.684989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.684989] env[61911]: _ensure_no_port_binding_failure(port) [ 547.684989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.684989] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.686277] env[61911]: nova.exception.PortBindingFailed: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. [ 547.686277] env[61911]: Removing descriptor: 16 [ 547.686277] env[61911]: ERROR nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Traceback (most recent call last): [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] yield resources [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self.driver.spawn(context, instance, image_meta, [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.686277] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] vm_ref = self.build_virtual_machine(instance, [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] for vif in network_info: [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return self._sync_wrapper(fn, *args, **kwargs) [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self.wait() [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self[:] = self._gt.wait() [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return self._exit_event.wait() [ 547.686908] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] result = hub.switch() [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return self.greenlet.switch() [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] result = function(*args, **kwargs) [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return func(*args, **kwargs) [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] raise e [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] nwinfo = self.network_api.allocate_for_instance( [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.687246] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] created_port_ids = self._update_ports_for_instance( [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] with excutils.save_and_reraise_exception(): [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self.force_reraise() [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] raise self.value [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] updated_port = self._update_port( [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] _ensure_no_port_binding_failure(port) [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.687596] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] raise exception.PortBindingFailed(port_id=port['id']) [ 547.687893] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] nova.exception.PortBindingFailed: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. [ 547.687893] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] [ 547.687893] env[61911]: INFO nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Terminating instance [ 547.765814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Releasing lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.766271] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 547.766454] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.767583] env[61911]: DEBUG oslo_concurrency.lockutils [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] Acquired lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.767765] env[61911]: DEBUG nova.network.neutron [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Refreshing network info cache for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 547.768785] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-303c3291-3f82-480a-9469-d99c618c7d91 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.785061] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e3aa47-bbaf-48f8-aefd-6a0004f1129d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.812496] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a5021d0d-325d-422f-a3db-eb51b129b72b could not be found. [ 547.812652] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.813116] env[61911]: INFO nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 547.813401] env[61911]: DEBUG oslo.service.loopingcall [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.813578] env[61911]: DEBUG nova.compute.manager [-] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 547.813669] env[61911]: DEBUG nova.network.neutron [-] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 547.875430] env[61911]: DEBUG nova.network.neutron [-] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.006855] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Successfully created port: 57872adc-9d66-49dd-a200-941b7beb6a75 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.156200] env[61911]: DEBUG nova.scheduler.client.report [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 548.190025] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Acquiring lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.190025] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Acquired lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.190025] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.331475] env[61911]: DEBUG nova.network.neutron [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.342911] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 548.366837] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 548.367236] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.367660] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 548.367660] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.367749] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 548.367895] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 548.368121] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 548.368301] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 548.368496] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 548.368631] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 548.369124] env[61911]: DEBUG nova.virt.hardware [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 548.370101] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e11803e-af4f-4855-9400-4ab3cef57fc9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.381611] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a9d5d3-e7b0-4c0a-b131-768ae67b9a62 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.387369] env[61911]: DEBUG nova.network.neutron [-] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.488061] env[61911]: DEBUG nova.network.neutron [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.667173] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.667173] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 548.674694] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.686s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.676699] env[61911]: INFO nova.compute.claims [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.735015] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.882647] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.891155] env[61911]: INFO nova.compute.manager [-] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Took 1.08 seconds to deallocate network for instance. [ 548.893841] env[61911]: DEBUG nova.compute.claims [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.894026] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.991761] env[61911]: DEBUG oslo_concurrency.lockutils [req-4b00defe-ea41-492c-9f42-f866565c0b54 req-af10dbb2-4673-45fd-a82a-00510a19e730 service nova] Releasing lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.177345] env[61911]: DEBUG nova.compute.utils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.178860] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 549.179432] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 549.385495] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Releasing lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.385923] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 549.386664] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 549.386664] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67b07ca9-9b46-4af1-aebd-7bbc3b1baa6e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.402147] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Acquiring lock "0ad2e485-4482-4514-845b-782c552d30fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.402576] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Lock "0ad2e485-4482-4514-845b-782c552d30fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.408281] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d803f177-9261-45b4-aec1-0141e940815a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.422915] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 549.438439] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 02ca2bcc-3a06-4e3d-9985-c158b541cdcc could not be found. [ 549.438825] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 549.439035] env[61911]: INFO nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 549.439377] env[61911]: DEBUG oslo.service.loopingcall [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 549.441029] env[61911]: DEBUG nova.compute.manager [-] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 549.441029] env[61911]: DEBUG nova.network.neutron [-] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.471425] env[61911]: DEBUG nova.network.neutron [-] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.522729] env[61911]: DEBUG nova.policy [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ee96b52414148b7bff1d067595848b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f40f512384d4fedb75bfc53f3195243', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 549.682607] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 549.722739] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Acquiring lock "38fcf50f-04c8-494b-9eb0-f62b25e200b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.724881] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Lock "38fcf50f-04c8-494b-9eb0-f62b25e200b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.907498] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c06488-25a2-45eb-b6e7-574d150780a2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.915874] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423e1dfb-5fb2-42a7-828c-b3849e889d6d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.953423] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fdc720-4d2f-418a-83e5-d0c7ea8f181f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.963345] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d755e0e-74e1-4e73-b43c-e98cdb3e1604 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.979850] env[61911]: DEBUG nova.network.neutron [-] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.982672] env[61911]: DEBUG nova.compute.provider_tree [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.983133] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.067021] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Successfully created port: 7fffd140-832b-417c-a31f-6d533c144442 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 550.459838] env[61911]: ERROR nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. [ 550.459838] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 550.459838] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.459838] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 550.459838] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.459838] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 550.459838] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.459838] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 550.459838] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.459838] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 550.459838] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.459838] env[61911]: ERROR nova.compute.manager raise self.value [ 550.459838] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.459838] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 550.459838] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.459838] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 550.460616] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.460616] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 550.460616] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. [ 550.460616] env[61911]: ERROR nova.compute.manager [ 550.460616] env[61911]: Traceback (most recent call last): [ 550.460616] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 550.460616] env[61911]: listener.cb(fileno) [ 550.460616] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.460616] env[61911]: result = function(*args, **kwargs) [ 550.460616] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.460616] env[61911]: return func(*args, **kwargs) [ 550.460616] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.460616] env[61911]: raise e [ 550.460616] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.460616] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 550.460616] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.460616] env[61911]: created_port_ids = self._update_ports_for_instance( [ 550.460616] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.460616] env[61911]: with excutils.save_and_reraise_exception(): [ 550.460616] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.460616] env[61911]: self.force_reraise() [ 550.460616] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.460616] env[61911]: raise self.value [ 550.460616] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.460616] env[61911]: updated_port = self._update_port( [ 550.460616] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.460616] env[61911]: _ensure_no_port_binding_failure(port) [ 550.460616] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.460616] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 550.461364] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. [ 550.461364] env[61911]: Removing descriptor: 15 [ 550.461364] env[61911]: ERROR nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Traceback (most recent call last): [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] yield resources [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self.driver.spawn(context, instance, image_meta, [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.461364] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] vm_ref = self.build_virtual_machine(instance, [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] for vif in network_info: [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return self._sync_wrapper(fn, *args, **kwargs) [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self.wait() [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self[:] = self._gt.wait() [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return self._exit_event.wait() [ 550.461687] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] result = hub.switch() [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return self.greenlet.switch() [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] result = function(*args, **kwargs) [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return func(*args, **kwargs) [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] raise e [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] nwinfo = self.network_api.allocate_for_instance( [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.462049] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] created_port_ids = self._update_ports_for_instance( [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] with excutils.save_and_reraise_exception(): [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self.force_reraise() [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] raise self.value [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] updated_port = self._update_port( [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] _ensure_no_port_binding_failure(port) [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.462371] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] raise exception.PortBindingFailed(port_id=port['id']) [ 550.462659] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] nova.exception.PortBindingFailed: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. [ 550.462659] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] [ 550.462659] env[61911]: INFO nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Terminating instance [ 550.484244] env[61911]: INFO nova.compute.manager [-] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Took 1.04 seconds to deallocate network for instance. [ 550.485083] env[61911]: DEBUG nova.scheduler.client.report [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 550.490817] env[61911]: DEBUG nova.compute.claims [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 550.490988] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.534421] env[61911]: DEBUG nova.compute.manager [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Received event network-vif-deleted-9df9710d-cc80-4012-a4a7-00c351ecd8f2 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 550.534787] env[61911]: DEBUG nova.compute.manager [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Received event network-changed-bddd1214-67ff-49f9-a0bb-e01cb3b8fc05 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 550.535220] env[61911]: DEBUG nova.compute.manager [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Refreshing instance network info cache due to event network-changed-bddd1214-67ff-49f9-a0bb-e01cb3b8fc05. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 550.535466] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] Acquiring lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.535643] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] Acquired lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.535937] env[61911]: DEBUG nova.network.neutron [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Refreshing network info cache for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 550.694635] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 550.718036] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 550.718278] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.718428] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 550.718612] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.718753] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 550.718893] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 550.719188] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 550.719371] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 550.719536] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 550.719721] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 550.719856] env[61911]: DEBUG nova.virt.hardware [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 550.720810] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffeb5cab-af8d-4285-ab29-74325dc84e8a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.732298] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dff36ea-ef99-40c0-9606-6007b9b5116d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.970464] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.971129] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquired lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.971129] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.991658] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.317s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.992565] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 550.995138] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.733s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.998916] env[61911]: INFO nova.compute.claims [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.069330] env[61911]: DEBUG nova.network.neutron [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.179279] env[61911]: DEBUG nova.network.neutron [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.503388] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.506709] env[61911]: DEBUG nova.compute.utils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.514654] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 551.514901] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 551.572975] env[61911]: DEBUG nova.policy [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0cf09dd11d64e14b61a61875ef12f5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4e4a478411447d59d15bf110b0d0c86', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.618245] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.682955] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] Releasing lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.682955] env[61911]: DEBUG nova.compute.manager [req-5b9a3326-de5a-40e7-9efc-24f316eeba86 req-85cbc14d-f926-4a18-9d0f-d45dab80cb5c service nova] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Received event network-vif-deleted-bddd1214-67ff-49f9-a0bb-e01cb3b8fc05 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 551.753445] env[61911]: ERROR nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. [ 551.753445] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.753445] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.753445] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.753445] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.753445] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.753445] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.753445] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.753445] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.753445] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 551.753445] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.753445] env[61911]: ERROR nova.compute.manager raise self.value [ 551.753445] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.753445] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.753445] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.753445] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.753974] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.753974] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.753974] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. [ 551.753974] env[61911]: ERROR nova.compute.manager [ 551.753974] env[61911]: Traceback (most recent call last): [ 551.753974] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.753974] env[61911]: listener.cb(fileno) [ 551.753974] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.753974] env[61911]: result = function(*args, **kwargs) [ 551.753974] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.753974] env[61911]: return func(*args, **kwargs) [ 551.753974] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 551.753974] env[61911]: raise e [ 551.753974] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.753974] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 551.753974] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.753974] env[61911]: created_port_ids = self._update_ports_for_instance( [ 551.753974] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.753974] env[61911]: with excutils.save_and_reraise_exception(): [ 551.753974] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.753974] env[61911]: self.force_reraise() [ 551.753974] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.753974] env[61911]: raise self.value [ 551.753974] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.753974] env[61911]: updated_port = self._update_port( [ 551.753974] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.753974] env[61911]: _ensure_no_port_binding_failure(port) [ 551.753974] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.753974] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.754761] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. [ 551.754761] env[61911]: Removing descriptor: 16 [ 551.754761] env[61911]: ERROR nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Traceback (most recent call last): [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] yield resources [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self.driver.spawn(context, instance, image_meta, [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.754761] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] vm_ref = self.build_virtual_machine(instance, [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] for vif in network_info: [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return self._sync_wrapper(fn, *args, **kwargs) [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self.wait() [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self[:] = self._gt.wait() [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return self._exit_event.wait() [ 551.759455] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] result = hub.switch() [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return self.greenlet.switch() [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] result = function(*args, **kwargs) [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return func(*args, **kwargs) [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] raise e [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] nwinfo = self.network_api.allocate_for_instance( [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.759809] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] created_port_ids = self._update_ports_for_instance( [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] with excutils.save_and_reraise_exception(): [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self.force_reraise() [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] raise self.value [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] updated_port = self._update_port( [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] _ensure_no_port_binding_failure(port) [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.760189] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] raise exception.PortBindingFailed(port_id=port['id']) [ 551.760580] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] nova.exception.PortBindingFailed: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. [ 551.760580] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] [ 551.760580] env[61911]: INFO nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Terminating instance [ 551.865360] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Acquiring lock "9f1af494-bb6c-4f7e-8218-0413848ceb5c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.865360] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Lock "9f1af494-bb6c-4f7e-8218-0413848ceb5c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.920570] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Successfully created port: 916ecd84-a31f-42d2-bfba-6253c4382ec9 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.016144] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 552.119764] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Releasing lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.120232] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 552.120498] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.120794] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70010715-1344-4430-bee1-f88b89e015fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.133169] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1252830-32fd-4d5b-a339-f23bc8efebcb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.162214] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 54cd80a2-17fb-446d-b25e-d6d68f8bd538 could not be found. [ 552.162377] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.162561] env[61911]: INFO nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Took 0.04 seconds to destroy the instance on the hypervisor. [ 552.162812] env[61911]: DEBUG oslo.service.loopingcall [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.163336] env[61911]: DEBUG nova.compute.manager [-] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 552.163336] env[61911]: DEBUG nova.network.neutron [-] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.211325] env[61911]: DEBUG nova.network.neutron [-] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.231027] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52414cb-26f3-4fb0-8e6f-cd7b99d3f42f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.239192] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f865e7-9525-4637-9311-c0ff12d5523f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.281092] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.281092] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.281092] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.283764] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206adc51-5b34-44ec-8595-6438dca12bf4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.293441] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38acc5b-3319-4118-b6dd-5fa8eca3f3c1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.316112] env[61911]: DEBUG nova.compute.provider_tree [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.713994] env[61911]: DEBUG nova.network.neutron [-] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.806834] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.818840] env[61911]: DEBUG nova.scheduler.client.report [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 552.970030] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.030857] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 553.057493] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 553.057832] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.057994] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 553.058197] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.058345] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 553.058517] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 553.058731] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 553.058888] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 553.059064] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 553.059307] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 553.059887] env[61911]: DEBUG nova.virt.hardware [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 553.060763] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c661d1-d914-42f3-912a-66702f3ae1f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.070248] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd5458a-6ad3-4abb-a2ab-11ff5aab9632 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.218549] env[61911]: INFO nova.compute.manager [-] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Took 1.05 seconds to deallocate network for instance. [ 553.219882] env[61911]: DEBUG nova.compute.claims [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 553.220094] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.258431] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.258813] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.259037] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 553.259159] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Rebuilding the list of instances to heal {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 553.325310] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.326137] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 553.330067] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.993s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.332970] env[61911]: INFO nova.compute.claims [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 553.473653] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.474712] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 553.474712] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.474712] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfa4040d-3001-419b-ab97-34741566bbf4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.484875] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69539d5c-c6f8-45f0-a030-2c1d2ca9e138 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.513256] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 82e22142-4083-4716-95a9-dea2d0faf998 could not be found. [ 553.513545] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 553.517019] env[61911]: INFO nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Took 0.04 seconds to destroy the instance on the hypervisor. [ 553.517019] env[61911]: DEBUG oslo.service.loopingcall [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.517019] env[61911]: DEBUG nova.compute.manager [-] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 553.517019] env[61911]: DEBUG nova.network.neutron [-] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.540495] env[61911]: DEBUG nova.network.neutron [-] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.610472] env[61911]: ERROR nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. [ 553.610472] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.610472] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.610472] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.610472] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.610472] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.610472] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.610472] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.610472] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.610472] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 553.610472] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.610472] env[61911]: ERROR nova.compute.manager raise self.value [ 553.610472] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.610472] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.610472] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.610472] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.611215] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.611215] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.611215] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. [ 553.611215] env[61911]: ERROR nova.compute.manager [ 553.611215] env[61911]: Traceback (most recent call last): [ 553.611215] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.611215] env[61911]: listener.cb(fileno) [ 553.611215] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.611215] env[61911]: result = function(*args, **kwargs) [ 553.611215] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.611215] env[61911]: return func(*args, **kwargs) [ 553.611215] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 553.611215] env[61911]: raise e [ 553.611215] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.611215] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 553.611215] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.611215] env[61911]: created_port_ids = self._update_ports_for_instance( [ 553.611215] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.611215] env[61911]: with excutils.save_and_reraise_exception(): [ 553.611215] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.611215] env[61911]: self.force_reraise() [ 553.611215] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.611215] env[61911]: raise self.value [ 553.611215] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.611215] env[61911]: updated_port = self._update_port( [ 553.611215] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.611215] env[61911]: _ensure_no_port_binding_failure(port) [ 553.611215] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.611215] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.612026] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. [ 553.612026] env[61911]: Removing descriptor: 17 [ 553.612026] env[61911]: ERROR nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Traceback (most recent call last): [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] yield resources [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self.driver.spawn(context, instance, image_meta, [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.612026] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] vm_ref = self.build_virtual_machine(instance, [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] for vif in network_info: [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return self._sync_wrapper(fn, *args, **kwargs) [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self.wait() [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self[:] = self._gt.wait() [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return self._exit_event.wait() [ 553.612311] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] result = hub.switch() [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return self.greenlet.switch() [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] result = function(*args, **kwargs) [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return func(*args, **kwargs) [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] raise e [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] nwinfo = self.network_api.allocate_for_instance( [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.612615] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] created_port_ids = self._update_ports_for_instance( [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] with excutils.save_and_reraise_exception(): [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self.force_reraise() [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] raise self.value [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] updated_port = self._update_port( [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] _ensure_no_port_binding_failure(port) [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.612997] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] raise exception.PortBindingFailed(port_id=port['id']) [ 553.613335] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] nova.exception.PortBindingFailed: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. [ 553.613335] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] [ 553.613335] env[61911]: INFO nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Terminating instance [ 553.623470] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "821d3f6f-e847-4ddb-ac00-ea55af302383" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.623470] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "821d3f6f-e847-4ddb-ac00-ea55af302383" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.764384] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 553.764384] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 553.765623] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 553.765823] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 553.766014] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 553.766213] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 553.766330] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Didn't find any instances for network info cache update. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 553.766593] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.766901] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.767102] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.768425] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.768425] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.768425] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.824335] env[61911]: DEBUG nova.compute.manager [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Received event network-changed-57872adc-9d66-49dd-a200-941b7beb6a75 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 553.824335] env[61911]: DEBUG nova.compute.manager [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Refreshing instance network info cache due to event network-changed-57872adc-9d66-49dd-a200-941b7beb6a75. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 553.824335] env[61911]: DEBUG oslo_concurrency.lockutils [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] Acquiring lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.824335] env[61911]: DEBUG oslo_concurrency.lockutils [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] Acquired lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.824335] env[61911]: DEBUG nova.network.neutron [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Refreshing network info cache for port 57872adc-9d66-49dd-a200-941b7beb6a75 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 553.841536] env[61911]: DEBUG nova.compute.utils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.845848] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 553.846044] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 553.964510] env[61911]: DEBUG nova.policy [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97d3138b660d4a9ca8a68a3789977991', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae8dc06260e540aeaabcfd82050e1373', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.044194] env[61911]: DEBUG nova.network.neutron [-] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.116943] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Acquiring lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.117143] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Acquired lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.117322] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 554.273091] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Getting list of instances from cluster (obj){ [ 554.273091] env[61911]: value = "domain-c8" [ 554.273091] env[61911]: _type = "ClusterComputeResource" [ 554.273091] env[61911]: } {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 554.274249] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e2b725-6ba6-43db-961e-3aecb39f8f33 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.288379] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Got total of 0 instances {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 554.288555] env[61911]: WARNING nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] While synchronizing instance power states, found 7 instances in the database and 0 instances on the hypervisor. [ 554.288905] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid a5021d0d-325d-422f-a3db-eb51b129b72b {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 554.293021] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 02ca2bcc-3a06-4e3d-9985-c158b541cdcc {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 554.293021] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 54cd80a2-17fb-446d-b25e-d6d68f8bd538 {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 554.293021] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 82e22142-4083-4716-95a9-dea2d0faf998 {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 554.293021] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 554.293021] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66 {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 554.293021] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 960d5c25-2a54-4eb8-b175-27e56b9894d9 {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 554.293021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "a5021d0d-325d-422f-a3db-eb51b129b72b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.293713] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "02ca2bcc-3a06-4e3d-9985-c158b541cdcc" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.293713] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "54cd80a2-17fb-446d-b25e-d6d68f8bd538" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.293713] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "82e22142-4083-4716-95a9-dea2d0faf998" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.293713] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.293713] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.294142] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "960d5c25-2a54-4eb8-b175-27e56b9894d9" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.294142] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.294142] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 554.294142] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.307109] env[61911]: DEBUG nova.compute.manager [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Received event network-changed-916ecd84-a31f-42d2-bfba-6253c4382ec9 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 554.307348] env[61911]: DEBUG nova.compute.manager [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Refreshing instance network info cache due to event network-changed-916ecd84-a31f-42d2-bfba-6253c4382ec9. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 554.307535] env[61911]: DEBUG oslo_concurrency.lockutils [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] Acquiring lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.347656] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 554.363313] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Acquiring lock "2bb4e0d0-8154-45d8-b662-02e5a9c28b21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.363602] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Lock "2bb4e0d0-8154-45d8-b662-02e5a9c28b21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.387040] env[61911]: DEBUG nova.network.neutron [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.487245] env[61911]: DEBUG nova.network.neutron [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.549393] env[61911]: INFO nova.compute.manager [-] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Took 1.03 seconds to deallocate network for instance. [ 554.554971] env[61911]: DEBUG nova.compute.claims [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.555217] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.594785] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a02adc-1634-4dbb-80d1-706977a18731 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.603586] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e44c13a-5011-4d52-a780-1ea47a9343ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.643946] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38516ae3-d902-4c99-b298-1509b4fd1d74 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.652296] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cc3c64-8184-4613-b76e-8ac7193671cf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.667658] env[61911]: DEBUG nova.compute.provider_tree [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.670337] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.796531] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.824006] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.990477] env[61911]: DEBUG oslo_concurrency.lockutils [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] Releasing lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.990709] env[61911]: DEBUG nova.compute.manager [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Received event network-vif-deleted-57872adc-9d66-49dd-a200-941b7beb6a75 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 554.991270] env[61911]: DEBUG nova.compute.manager [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Received event network-changed-7fffd140-832b-417c-a31f-6d533c144442 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 554.992514] env[61911]: DEBUG nova.compute.manager [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Refreshing instance network info cache due to event network-changed-7fffd140-832b-417c-a31f-6d533c144442. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 554.992514] env[61911]: DEBUG oslo_concurrency.lockutils [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] Acquiring lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.992514] env[61911]: DEBUG oslo_concurrency.lockutils [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] Acquired lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.992514] env[61911]: DEBUG nova.network.neutron [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Refreshing network info cache for port 7fffd140-832b-417c-a31f-6d533c144442 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 555.174403] env[61911]: DEBUG nova.scheduler.client.report [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 555.296272] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Successfully created port: 29dbc5e9-60eb-47ed-b07f-c125a3587158 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.328452] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Releasing lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.328452] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 555.328452] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 555.328452] env[61911]: DEBUG oslo_concurrency.lockutils [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] Acquired lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.328452] env[61911]: DEBUG nova.network.neutron [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Refreshing network info cache for port 916ecd84-a31f-42d2-bfba-6253c4382ec9 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 555.331781] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-083b965a-4f6b-4c0e-9a22-824bd5d98f85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.342519] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3471e59d-c77a-4785-bf00-48bd8fe1671e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.367489] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 555.372212] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba could not be found. [ 555.372212] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 555.372339] env[61911]: INFO nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Took 0.04 seconds to destroy the instance on the hypervisor. [ 555.372601] env[61911]: DEBUG oslo.service.loopingcall [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.372819] env[61911]: DEBUG nova.compute.manager [-] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 555.373483] env[61911]: DEBUG nova.network.neutron [-] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 555.396804] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 555.396804] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.396804] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 555.397024] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.397062] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 555.398869] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 555.398869] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 555.398869] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 555.398869] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 555.398869] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 555.399117] env[61911]: DEBUG nova.virt.hardware [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 555.399117] env[61911]: DEBUG nova.network.neutron [-] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.400786] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226d6a0c-a2aa-476e-95a4-53ef80a39ab0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.409708] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7836fc-917f-4f89-8acc-8b1f1e0ae8c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.521313] env[61911]: DEBUG nova.network.neutron [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.610793] env[61911]: DEBUG nova.network.neutron [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.684295] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.684844] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 555.690180] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.583s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.691750] env[61911]: INFO nova.compute.claims [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.856978] env[61911]: DEBUG nova.network.neutron [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.868891] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.869191] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.904154] env[61911]: DEBUG nova.network.neutron [-] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.004317] env[61911]: DEBUG nova.network.neutron [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.113905] env[61911]: DEBUG oslo_concurrency.lockutils [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] Releasing lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.114185] env[61911]: DEBUG nova.compute.manager [req-2b6f7e50-57f9-4fd5-9ea8-6961e1696f68 req-4d5b7609-91cc-4fa9-9b2b-b146da9650d1 service nova] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Received event network-vif-deleted-7fffd140-832b-417c-a31f-6d533c144442 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 556.198644] env[61911]: DEBUG nova.compute.utils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 556.202896] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 556.203779] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 556.409483] env[61911]: INFO nova.compute.manager [-] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Took 1.03 seconds to deallocate network for instance. [ 556.412390] env[61911]: DEBUG nova.compute.claims [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.412390] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.413272] env[61911]: DEBUG nova.policy [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96721947abba4d33b8838e88fa88c387', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7d9df8368db43609a412e2ffed933d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 556.507499] env[61911]: DEBUG oslo_concurrency.lockutils [req-03aedd4d-86de-4420-9664-b3113682e4b9 req-b165fe38-aada-4a96-a8d7-7d85f04703e9 service nova] Releasing lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.712325] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 556.932640] env[61911]: DEBUG nova.compute.manager [req-9c446afd-c7e7-4da2-8f54-313a76727ae0 req-7888254e-85a5-449c-a10d-f6d5dc437b19 service nova] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Received event network-vif-deleted-916ecd84-a31f-42d2-bfba-6253c4382ec9 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 556.971098] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadfd0ef-b3f3-47e7-941c-97cc8ff90fb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.980692] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9a15b5-b151-4014-8fd5-49f6c55df8ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.018034] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad0fb27-4ac2-4dd0-8bea-864b47af3652 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.027990] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b371eb-adb1-4e82-a856-3446595eaf4a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.045156] env[61911]: DEBUG nova.compute.provider_tree [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.548726] env[61911]: DEBUG nova.scheduler.client.report [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 557.729149] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 557.776691] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 557.776691] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.776691] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 557.776691] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.776970] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 557.776970] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 557.776970] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 557.776970] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 557.776970] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 557.777136] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 557.777136] env[61911]: DEBUG nova.virt.hardware [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 557.778116] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645d4b13-6480-49a5-afae-67e5e50a18b0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.788572] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2136eb29-f9f7-4793-b542-3d7d28dee12e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.865923] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Successfully created port: 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.054805] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.055344] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 558.058864] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.310s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.060274] env[61911]: INFO nova.compute.claims [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.207176] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Acquiring lock "8da2407a-c7e5-437a-ab71-57b92d98182f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.207670] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Lock "8da2407a-c7e5-437a-ab71-57b92d98182f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.569891] env[61911]: DEBUG nova.compute.utils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 558.580021] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 558.580177] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 558.695829] env[61911]: DEBUG nova.policy [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f76b9db552fc4bc6b3eb44de6e7bec58', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcfc4f0fd300417e8cad41fad828fd5d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.081800] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 559.308690] env[61911]: DEBUG nova.compute.manager [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Received event network-changed-29dbc5e9-60eb-47ed-b07f-c125a3587158 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 559.308690] env[61911]: DEBUG nova.compute.manager [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Refreshing instance network info cache due to event network-changed-29dbc5e9-60eb-47ed-b07f-c125a3587158. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 559.308690] env[61911]: DEBUG oslo_concurrency.lockutils [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] Acquiring lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.309515] env[61911]: DEBUG oslo_concurrency.lockutils [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] Acquired lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.310191] env[61911]: DEBUG nova.network.neutron [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Refreshing network info cache for port 29dbc5e9-60eb-47ed-b07f-c125a3587158 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 559.404641] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd4fab9-fbb3-40ad-88b8-98e0994ca4f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.412993] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106a3d04-f695-4101-be3a-c2764a9c0d9f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.453885] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabef076-05c8-4d40-990a-c69c8b9264b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.462716] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53969deb-0300-476d-8333-f57949a9cb45 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.477537] env[61911]: DEBUG nova.compute.provider_tree [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.772069] env[61911]: ERROR nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. [ 559.772069] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.772069] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 559.772069] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.772069] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.772069] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.772069] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.772069] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.772069] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.772069] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 559.772069] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.772069] env[61911]: ERROR nova.compute.manager raise self.value [ 559.772069] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.772069] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.772069] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.772069] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.772689] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.772689] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.772689] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. [ 559.772689] env[61911]: ERROR nova.compute.manager [ 559.772689] env[61911]: Traceback (most recent call last): [ 559.772689] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.772689] env[61911]: listener.cb(fileno) [ 559.772689] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.772689] env[61911]: result = function(*args, **kwargs) [ 559.772689] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.772689] env[61911]: return func(*args, **kwargs) [ 559.772689] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 559.772689] env[61911]: raise e [ 559.772689] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 559.772689] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 559.772689] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.772689] env[61911]: created_port_ids = self._update_ports_for_instance( [ 559.772689] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.772689] env[61911]: with excutils.save_and_reraise_exception(): [ 559.772689] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.772689] env[61911]: self.force_reraise() [ 559.772689] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.772689] env[61911]: raise self.value [ 559.772689] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.772689] env[61911]: updated_port = self._update_port( [ 559.772689] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.772689] env[61911]: _ensure_no_port_binding_failure(port) [ 559.772689] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.772689] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.773988] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. [ 559.773988] env[61911]: Removing descriptor: 17 [ 559.773988] env[61911]: ERROR nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Traceback (most recent call last): [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] yield resources [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self.driver.spawn(context, instance, image_meta, [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.773988] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] vm_ref = self.build_virtual_machine(instance, [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] for vif in network_info: [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return self._sync_wrapper(fn, *args, **kwargs) [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self.wait() [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self[:] = self._gt.wait() [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return self._exit_event.wait() [ 559.774496] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] result = hub.switch() [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return self.greenlet.switch() [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] result = function(*args, **kwargs) [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return func(*args, **kwargs) [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] raise e [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] nwinfo = self.network_api.allocate_for_instance( [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.774840] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] created_port_ids = self._update_ports_for_instance( [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] with excutils.save_and_reraise_exception(): [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self.force_reraise() [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] raise self.value [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] updated_port = self._update_port( [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] _ensure_no_port_binding_failure(port) [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.776147] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] raise exception.PortBindingFailed(port_id=port['id']) [ 559.776450] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] nova.exception.PortBindingFailed: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. [ 559.776450] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] [ 559.776450] env[61911]: INFO nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Terminating instance [ 559.858160] env[61911]: DEBUG nova.network.neutron [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.983837] env[61911]: DEBUG nova.scheduler.client.report [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 560.097228] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 560.124568] env[61911]: DEBUG nova.network.neutron [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.146624] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 560.146624] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.146889] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 560.146889] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.148218] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 560.148218] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 560.148971] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 560.148971] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 560.148971] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 560.148971] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 560.148971] env[61911]: DEBUG nova.virt.hardware [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 560.150301] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b5ad70-4b1c-4c25-8a7c-38c7bf3a9c3c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.161931] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72ef16e-4141-4e53-995b-4d11d815bec2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.278888] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Acquiring lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.354148] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Successfully created port: 54603ce5-e295-4258-8877-4387d9f81d28 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 560.490089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.490693] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 560.494529] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.600s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.628556] env[61911]: DEBUG oslo_concurrency.lockutils [req-d23eadbf-3b0b-4d41-ba40-7c393ba0708e req-1f89c94e-9804-4826-bcf9-5210ae77d2ed service nova] Releasing lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.629285] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Acquired lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.629285] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.999225] env[61911]: DEBUG nova.compute.utils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.007153] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 561.007153] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 561.181079] env[61911]: DEBUG nova.policy [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f585bde44cde40bcbff7f8262c4a01af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113a8c86019d40d483097b9999561030', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 561.185371] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.314350] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7e8fb7-066b-4cd1-bbe6-ee417f385823 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.324043] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cfa779-89d6-4304-9570-421cabd2e333 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.355953] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea013aa-d882-4f1c-9051-24c5d8de33bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.364836] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff25ca8f-2593-44b4-98b9-224db8823b1d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.383922] env[61911]: DEBUG nova.compute.provider_tree [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.506339] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 561.630269] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.890369] env[61911]: DEBUG nova.scheduler.client.report [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 562.139355] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Releasing lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.139355] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 562.139355] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 562.139355] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fde340d1-889d-4e19-ae2b-8fc48db172e5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.152239] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e232bc3-e3d4-43a9-a69d-dece4cc802c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.180655] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66 could not be found. [ 562.181105] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 562.181769] env[61911]: INFO nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Took 0.04 seconds to destroy the instance on the hypervisor. [ 562.182208] env[61911]: DEBUG oslo.service.loopingcall [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.182584] env[61911]: DEBUG nova.compute.manager [-] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 562.182891] env[61911]: DEBUG nova.network.neutron [-] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.199146] env[61911]: DEBUG nova.compute.manager [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Received event network-changed-80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 562.199146] env[61911]: DEBUG nova.compute.manager [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Refreshing instance network info cache due to event network-changed-80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 562.199146] env[61911]: DEBUG oslo_concurrency.lockutils [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] Acquiring lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.199146] env[61911]: DEBUG oslo_concurrency.lockutils [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] Acquired lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.199146] env[61911]: DEBUG nova.network.neutron [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Refreshing network info cache for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 562.218425] env[61911]: DEBUG nova.compute.manager [req-00d8dc89-2337-438c-bf36-e69f98c674a9 req-06d7aefb-8c6e-4664-bbd8-fa3b0e3aaa89 service nova] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Received event network-vif-deleted-29dbc5e9-60eb-47ed-b07f-c125a3587158 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 562.252287] env[61911]: DEBUG nova.network.neutron [-] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.400999] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.401723] env[61911]: ERROR nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Traceback (most recent call last): [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self.driver.spawn(context, instance, image_meta, [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] vm_ref = self.build_virtual_machine(instance, [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.401723] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] for vif in network_info: [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return self._sync_wrapper(fn, *args, **kwargs) [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self.wait() [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self[:] = self._gt.wait() [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return self._exit_event.wait() [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] result = hub.switch() [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.402117] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return self.greenlet.switch() [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] result = function(*args, **kwargs) [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] return func(*args, **kwargs) [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] raise e [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] nwinfo = self.network_api.allocate_for_instance( [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] created_port_ids = self._update_ports_for_instance( [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] with excutils.save_and_reraise_exception(): [ 562.402739] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] self.force_reraise() [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] raise self.value [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] updated_port = self._update_port( [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] _ensure_no_port_binding_failure(port) [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] raise exception.PortBindingFailed(port_id=port['id']) [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] nova.exception.PortBindingFailed: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. [ 562.403072] env[61911]: ERROR nova.compute.manager [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] [ 562.403346] env[61911]: DEBUG nova.compute.utils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 562.404396] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.421s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.406297] env[61911]: INFO nova.compute.claims [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.411602] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Build of instance a5021d0d-325d-422f-a3db-eb51b129b72b was re-scheduled: Binding failed for port 9df9710d-cc80-4012-a4a7-00c351ecd8f2, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 562.412115] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 562.412481] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Acquiring lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.412783] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Acquired lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.413040] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.515328] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 562.547022] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 562.547022] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.547022] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 562.547426] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.547426] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 562.547426] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 562.547426] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 562.547426] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 562.547665] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 562.547665] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 562.549363] env[61911]: DEBUG nova.virt.hardware [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 562.550668] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e065b3a1-5aad-4567-b5ce-f1be28bce105 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.558742] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d1e27b-1196-4e40-bb8f-9610caf78c08 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.755031] env[61911]: DEBUG nova.network.neutron [-] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.802196] env[61911]: DEBUG nova.network.neutron [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.876844] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Successfully created port: f419ee8d-147a-45b7-bf04-96850a6effa7 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.901341] env[61911]: ERROR nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. [ 562.901341] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 562.901341] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 562.901341] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 562.901341] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.901341] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 562.901341] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.901341] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 562.901341] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.901341] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 562.901341] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.901341] env[61911]: ERROR nova.compute.manager raise self.value [ 562.901341] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.901341] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 562.901341] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.901341] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 562.901825] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.901825] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 562.901825] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. [ 562.901825] env[61911]: ERROR nova.compute.manager [ 562.901825] env[61911]: Traceback (most recent call last): [ 562.901825] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 562.901825] env[61911]: listener.cb(fileno) [ 562.901825] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.901825] env[61911]: result = function(*args, **kwargs) [ 562.901825] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.901825] env[61911]: return func(*args, **kwargs) [ 562.901825] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 562.901825] env[61911]: raise e [ 562.901825] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 562.901825] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 562.901825] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.901825] env[61911]: created_port_ids = self._update_ports_for_instance( [ 562.901825] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.901825] env[61911]: with excutils.save_and_reraise_exception(): [ 562.901825] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.901825] env[61911]: self.force_reraise() [ 562.901825] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.901825] env[61911]: raise self.value [ 562.901825] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.901825] env[61911]: updated_port = self._update_port( [ 562.901825] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.901825] env[61911]: _ensure_no_port_binding_failure(port) [ 562.901825] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.901825] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 562.902744] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. [ 562.902744] env[61911]: Removing descriptor: 16 [ 562.902744] env[61911]: ERROR nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Traceback (most recent call last): [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] yield resources [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self.driver.spawn(context, instance, image_meta, [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.902744] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] vm_ref = self.build_virtual_machine(instance, [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] for vif in network_info: [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return self._sync_wrapper(fn, *args, **kwargs) [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self.wait() [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self[:] = self._gt.wait() [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return self._exit_event.wait() [ 562.903072] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] result = hub.switch() [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return self.greenlet.switch() [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] result = function(*args, **kwargs) [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return func(*args, **kwargs) [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] raise e [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] nwinfo = self.network_api.allocate_for_instance( [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.903461] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] created_port_ids = self._update_ports_for_instance( [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] with excutils.save_and_reraise_exception(): [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self.force_reraise() [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] raise self.value [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] updated_port = self._update_port( [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] _ensure_no_port_binding_failure(port) [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.903830] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] raise exception.PortBindingFailed(port_id=port['id']) [ 562.904233] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] nova.exception.PortBindingFailed: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. [ 562.904233] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] [ 562.904233] env[61911]: INFO nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Terminating instance [ 562.976946] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.075446] env[61911]: DEBUG nova.network.neutron [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.259471] env[61911]: INFO nova.compute.manager [-] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Took 1.07 seconds to deallocate network for instance. [ 563.261427] env[61911]: DEBUG nova.compute.claims [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 563.262075] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.315225] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.408067] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.577143] env[61911]: DEBUG oslo_concurrency.lockutils [req-35fa2a31-ee6d-4ed8-ac68-194ac6b3f832 req-79fadcdd-8ada-4839-b17a-b0fd979fe73f service nova] Releasing lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.577518] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquired lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.577851] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.691283] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa5fc3a-237b-4316-b4c3-d59021df171b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.701527] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eef9772-48f0-4907-95b0-df59d13e1c45 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.737743] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491117b6-b498-4737-a94a-8b7c94ed4902 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.750921] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601d4c3e-e0d7-41ee-8b21-f45c6c03052f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.767864] env[61911]: DEBUG nova.compute.provider_tree [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.818962] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Releasing lock "refresh_cache-a5021d0d-325d-422f-a3db-eb51b129b72b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.819444] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 563.819715] env[61911]: DEBUG nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 563.820072] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.908064] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.183387] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.275081] env[61911]: DEBUG nova.scheduler.client.report [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 564.411727] env[61911]: DEBUG nova.network.neutron [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.549334] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.619063] env[61911]: DEBUG nova.compute.manager [req-56f26de1-a9c6-47e0-8202-00cfefe12045 req-3d0a6119-4594-4b20-8fc8-cc52f77c3f3e service nova] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Received event network-vif-deleted-80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 564.780540] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.780957] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 564.787677] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.297s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.915298] env[61911]: INFO nova.compute.manager [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] Took 1.10 seconds to deallocate network for instance. [ 565.053854] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Releasing lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.054664] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 565.055109] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.055559] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f916bc6f-0850-47f6-9ee7-3e605672b7bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.069617] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ef3bac-d530-44c3-b074-133d6d02e890 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.095805] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 960d5c25-2a54-4eb8-b175-27e56b9894d9 could not be found. [ 565.098019] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.098019] env[61911]: INFO nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 565.098019] env[61911]: DEBUG oslo.service.loopingcall [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.098019] env[61911]: DEBUG nova.compute.manager [-] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 565.098019] env[61911]: DEBUG nova.network.neutron [-] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.159732] env[61911]: DEBUG nova.network.neutron [-] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.299018] env[61911]: DEBUG nova.compute.utils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 565.307213] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 565.307213] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 565.656622] env[61911]: DEBUG nova.policy [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ab22c56923b4706b6ed6dca24cd3206', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ba0abf0877a45bcaba6879e1a7fadff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 565.662418] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd07d9af-f359-4580-81c6-0d5c064d1324 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.665852] env[61911]: DEBUG nova.network.neutron [-] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.673975] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a28a5b2-9745-490e-9fad-7a6fc9a70ea1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.718025] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d865117f-9df4-46cb-9077-928da68b935a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.737045] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1bda40-a065-4cff-b055-e19b9d889814 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.757604] env[61911]: DEBUG nova.compute.provider_tree [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.805253] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 565.973854] env[61911]: INFO nova.scheduler.client.report [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Deleted allocations for instance a5021d0d-325d-422f-a3db-eb51b129b72b [ 566.172448] env[61911]: INFO nova.compute.manager [-] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Took 1.08 seconds to deallocate network for instance. [ 566.176091] env[61911]: DEBUG nova.compute.claims [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.176278] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.263661] env[61911]: DEBUG nova.scheduler.client.report [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 566.489704] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d604ffa1-88d3-4f62-bc7b-0a123ab517a0 tempest-ServerDiagnosticsTest-1680918676 tempest-ServerDiagnosticsTest-1680918676-project-member] Lock "a5021d0d-325d-422f-a3db-eb51b129b72b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.075s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.492735] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "a5021d0d-325d-422f-a3db-eb51b129b72b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 12.202s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.492940] env[61911]: INFO nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: a5021d0d-325d-422f-a3db-eb51b129b72b] During sync_power_state the instance has a pending task (spawning). Skip. [ 566.495029] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "a5021d0d-325d-422f-a3db-eb51b129b72b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.774930] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.987s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.775588] env[61911]: ERROR nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Traceback (most recent call last): [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self.driver.spawn(context, instance, image_meta, [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] vm_ref = self.build_virtual_machine(instance, [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.775588] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] for vif in network_info: [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return self._sync_wrapper(fn, *args, **kwargs) [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self.wait() [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self[:] = self._gt.wait() [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return self._exit_event.wait() [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] result = hub.switch() [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.775958] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return self.greenlet.switch() [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] result = function(*args, **kwargs) [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] return func(*args, **kwargs) [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] raise e [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] nwinfo = self.network_api.allocate_for_instance( [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] created_port_ids = self._update_ports_for_instance( [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] with excutils.save_and_reraise_exception(): [ 566.776404] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] self.force_reraise() [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] raise self.value [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] updated_port = self._update_port( [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] _ensure_no_port_binding_failure(port) [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] raise exception.PortBindingFailed(port_id=port['id']) [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] nova.exception.PortBindingFailed: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. [ 566.776775] env[61911]: ERROR nova.compute.manager [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] [ 566.777107] env[61911]: DEBUG nova.compute.utils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.778668] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.558s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.782136] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Build of instance 02ca2bcc-3a06-4e3d-9985-c158b541cdcc was re-scheduled: Binding failed for port bddd1214-67ff-49f9-a0bb-e01cb3b8fc05, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 566.782441] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 566.782699] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Acquiring lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.782873] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Acquired lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.783132] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 566.819679] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 566.857794] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 566.858568] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.858709] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 566.859060] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.859914] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 566.859914] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 566.859914] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 566.861329] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 566.861329] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 566.861329] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 566.861329] env[61911]: DEBUG nova.virt.hardware [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 566.862558] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b027149-8cdb-4c3b-9724-8eb458adba1c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.872775] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29355bfd-9fb3-4e24-8aaf-bfec97c0724f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.992628] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 567.387307] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.524963] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.557032] env[61911]: DEBUG nova.compute.manager [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Received event network-changed-54603ce5-e295-4258-8877-4387d9f81d28 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 567.557206] env[61911]: DEBUG nova.compute.manager [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Refreshing instance network info cache due to event network-changed-54603ce5-e295-4258-8877-4387d9f81d28. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 567.557486] env[61911]: DEBUG oslo_concurrency.lockutils [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] Acquiring lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.557892] env[61911]: DEBUG oslo_concurrency.lockutils [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] Acquired lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.557892] env[61911]: DEBUG nova.network.neutron [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Refreshing network info cache for port 54603ce5-e295-4258-8877-4387d9f81d28 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.600993] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9c082f-b85b-425f-bdb8-1e8b40dc9422 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.609109] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76846015-a197-4cad-9131-d19e57edbdd7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.648751] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7cc82a-6715-463b-8635-965c897c853a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.656597] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125fb990-0201-4ac8-bc82-9adeb761dcaa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.672236] env[61911]: DEBUG nova.compute.provider_tree [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.709568] env[61911]: DEBUG nova.compute.manager [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Received event network-changed-f419ee8d-147a-45b7-bf04-96850a6effa7 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 567.709568] env[61911]: DEBUG nova.compute.manager [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Refreshing instance network info cache due to event network-changed-f419ee8d-147a-45b7-bf04-96850a6effa7. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 567.709813] env[61911]: DEBUG oslo_concurrency.lockutils [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] Acquiring lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.709964] env[61911]: DEBUG oslo_concurrency.lockutils [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] Acquired lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.710455] env[61911]: DEBUG nova.network.neutron [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Refreshing network info cache for port f419ee8d-147a-45b7-bf04-96850a6effa7 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 567.738581] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.175957] env[61911]: DEBUG nova.scheduler.client.report [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 568.241932] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Releasing lock "refresh_cache-02ca2bcc-3a06-4e3d-9985-c158b541cdcc" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.241932] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 568.241932] env[61911]: DEBUG nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 568.241932] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.268711] env[61911]: ERROR nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. [ 568.268711] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.268711] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 568.268711] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.268711] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.268711] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.268711] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.268711] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.268711] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.268711] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 568.268711] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.268711] env[61911]: ERROR nova.compute.manager raise self.value [ 568.268711] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.268711] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.268711] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.268711] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.269275] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.269275] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.269275] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. [ 568.269275] env[61911]: ERROR nova.compute.manager [ 568.269275] env[61911]: Traceback (most recent call last): [ 568.269275] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.269275] env[61911]: listener.cb(fileno) [ 568.269275] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.269275] env[61911]: result = function(*args, **kwargs) [ 568.269275] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.269275] env[61911]: return func(*args, **kwargs) [ 568.269275] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 568.269275] env[61911]: raise e [ 568.269275] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 568.269275] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 568.269275] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.269275] env[61911]: created_port_ids = self._update_ports_for_instance( [ 568.269275] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.269275] env[61911]: with excutils.save_and_reraise_exception(): [ 568.269275] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.269275] env[61911]: self.force_reraise() [ 568.269275] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.269275] env[61911]: raise self.value [ 568.269275] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.269275] env[61911]: updated_port = self._update_port( [ 568.269275] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.269275] env[61911]: _ensure_no_port_binding_failure(port) [ 568.269275] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.269275] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.270035] env[61911]: nova.exception.PortBindingFailed: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. [ 568.270035] env[61911]: Removing descriptor: 17 [ 568.270035] env[61911]: ERROR nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Traceback (most recent call last): [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] yield resources [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self.driver.spawn(context, instance, image_meta, [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.270035] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] vm_ref = self.build_virtual_machine(instance, [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] for vif in network_info: [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return self._sync_wrapper(fn, *args, **kwargs) [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self.wait() [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self[:] = self._gt.wait() [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return self._exit_event.wait() [ 568.270409] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] result = hub.switch() [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return self.greenlet.switch() [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] result = function(*args, **kwargs) [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return func(*args, **kwargs) [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] raise e [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] nwinfo = self.network_api.allocate_for_instance( [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.270730] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] created_port_ids = self._update_ports_for_instance( [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] with excutils.save_and_reraise_exception(): [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self.force_reraise() [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] raise self.value [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] updated_port = self._update_port( [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] _ensure_no_port_binding_failure(port) [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.271044] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] raise exception.PortBindingFailed(port_id=port['id']) [ 568.271406] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] nova.exception.PortBindingFailed: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. [ 568.271406] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] [ 568.271406] env[61911]: INFO nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Terminating instance [ 568.300594] env[61911]: DEBUG nova.network.neutron [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.303316] env[61911]: DEBUG nova.network.neutron [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.314205] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.331266] env[61911]: ERROR nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. [ 568.331266] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.331266] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 568.331266] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.331266] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.331266] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.331266] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.331266] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.331266] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.331266] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 568.331266] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.331266] env[61911]: ERROR nova.compute.manager raise self.value [ 568.331266] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.331266] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.331266] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.331266] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.332221] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.332221] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.332221] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. [ 568.332221] env[61911]: ERROR nova.compute.manager [ 568.332221] env[61911]: Traceback (most recent call last): [ 568.332221] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.332221] env[61911]: listener.cb(fileno) [ 568.332221] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.332221] env[61911]: result = function(*args, **kwargs) [ 568.332221] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.332221] env[61911]: return func(*args, **kwargs) [ 568.332221] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 568.332221] env[61911]: raise e [ 568.332221] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 568.332221] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 568.332221] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.332221] env[61911]: created_port_ids = self._update_ports_for_instance( [ 568.332221] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.332221] env[61911]: with excutils.save_and_reraise_exception(): [ 568.332221] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.332221] env[61911]: self.force_reraise() [ 568.332221] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.332221] env[61911]: raise self.value [ 568.332221] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.332221] env[61911]: updated_port = self._update_port( [ 568.332221] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.332221] env[61911]: _ensure_no_port_binding_failure(port) [ 568.332221] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.332221] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.333041] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. [ 568.333041] env[61911]: Removing descriptor: 15 [ 568.333100] env[61911]: ERROR nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Traceback (most recent call last): [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] yield resources [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self.driver.spawn(context, instance, image_meta, [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] vm_ref = self.build_virtual_machine(instance, [ 568.333100] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] for vif in network_info: [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return self._sync_wrapper(fn, *args, **kwargs) [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self.wait() [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self[:] = self._gt.wait() [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return self._exit_event.wait() [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.333380] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] result = hub.switch() [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return self.greenlet.switch() [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] result = function(*args, **kwargs) [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return func(*args, **kwargs) [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] raise e [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] nwinfo = self.network_api.allocate_for_instance( [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] created_port_ids = self._update_ports_for_instance( [ 568.333703] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] with excutils.save_and_reraise_exception(): [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self.force_reraise() [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] raise self.value [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] updated_port = self._update_port( [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] _ensure_no_port_binding_failure(port) [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] raise exception.PortBindingFailed(port_id=port['id']) [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] nova.exception.PortBindingFailed: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. [ 568.334024] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] [ 568.334422] env[61911]: INFO nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Terminating instance [ 568.482237] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Successfully created port: 43e0a58a-e1c5-4f2a-a643-52815b9757d1 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.678236] env[61911]: DEBUG nova.network.neutron [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.689657] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.691839] env[61911]: ERROR nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Traceback (most recent call last): [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self.driver.spawn(context, instance, image_meta, [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] vm_ref = self.build_virtual_machine(instance, [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.691839] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] for vif in network_info: [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return self._sync_wrapper(fn, *args, **kwargs) [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self.wait() [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self[:] = self._gt.wait() [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return self._exit_event.wait() [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] result = hub.switch() [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.692482] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return self.greenlet.switch() [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] result = function(*args, **kwargs) [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] return func(*args, **kwargs) [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] raise e [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] nwinfo = self.network_api.allocate_for_instance( [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] created_port_ids = self._update_ports_for_instance( [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] with excutils.save_and_reraise_exception(): [ 568.692872] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] self.force_reraise() [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] raise self.value [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] updated_port = self._update_port( [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] _ensure_no_port_binding_failure(port) [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] raise exception.PortBindingFailed(port_id=port['id']) [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] nova.exception.PortBindingFailed: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. [ 568.693224] env[61911]: ERROR nova.compute.manager [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] [ 568.693518] env[61911]: DEBUG nova.compute.utils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 568.693518] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.137s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.697624] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Build of instance 54cd80a2-17fb-446d-b25e-d6d68f8bd538 was re-scheduled: Binding failed for port 57872adc-9d66-49dd-a200-941b7beb6a75, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 568.697624] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 568.697624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.697624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquired lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.697998] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 568.776542] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.822024] env[61911]: DEBUG nova.network.neutron [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.840641] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Acquiring lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.853524] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.853738] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.020145] env[61911]: DEBUG nova.network.neutron [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.184923] env[61911]: DEBUG oslo_concurrency.lockutils [req-12463c96-8dce-403e-beb5-3ae30c515197 req-c1f78982-cbde-4446-8a3d-fdde4259bd0f service nova] Releasing lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.185368] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Acquired lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.185544] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.320348] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.323729] env[61911]: INFO nova.compute.manager [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] Took 1.08 seconds to deallocate network for instance. [ 569.510974] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd2202e-67cb-45ea-ab01-dc48a3763b53 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.521040] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be9d6de-3dc5-4ce5-87f2-0ea587830edd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.523651] env[61911]: DEBUG oslo_concurrency.lockutils [req-d1f8dd3b-96bf-4cff-8529-160667d3bc0f req-36801ba0-333d-4e20-ae62-dbbb78095cd8 service nova] Releasing lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.524677] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquired lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.526015] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.557680] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e476aae-9116-464f-b4d4-bd821c5758bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.567132] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebd3346-31c4-4a91-8916-91fb3e1cdf17 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.581604] env[61911]: DEBUG nova.compute.provider_tree [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.612651] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.766569] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.088096] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.090693] env[61911]: DEBUG nova.scheduler.client.report [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 570.115234] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Releasing lock "refresh_cache-54cd80a2-17fb-446d-b25e-d6d68f8bd538" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.115886] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 570.116517] env[61911]: DEBUG nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 570.116517] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.192270] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.216399] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.378617] env[61911]: INFO nova.scheduler.client.report [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Deleted allocations for instance 02ca2bcc-3a06-4e3d-9985-c158b541cdcc [ 570.600847] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.601543] env[61911]: ERROR nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Traceback (most recent call last): [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self.driver.spawn(context, instance, image_meta, [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] vm_ref = self.build_virtual_machine(instance, [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.601543] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] for vif in network_info: [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return self._sync_wrapper(fn, *args, **kwargs) [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self.wait() [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self[:] = self._gt.wait() [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return self._exit_event.wait() [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] result = hub.switch() [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.601968] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return self.greenlet.switch() [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] result = function(*args, **kwargs) [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] return func(*args, **kwargs) [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] raise e [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] nwinfo = self.network_api.allocate_for_instance( [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] created_port_ids = self._update_ports_for_instance( [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] with excutils.save_and_reraise_exception(): [ 570.602613] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] self.force_reraise() [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] raise self.value [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] updated_port = self._update_port( [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] _ensure_no_port_binding_failure(port) [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] raise exception.PortBindingFailed(port_id=port['id']) [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] nova.exception.PortBindingFailed: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. [ 570.602937] env[61911]: ERROR nova.compute.manager [instance: 82e22142-4083-4716-95a9-dea2d0faf998] [ 570.603218] env[61911]: DEBUG nova.compute.utils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 570.608369] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Releasing lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.608369] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 570.608488] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 570.609113] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Build of instance 82e22142-4083-4716-95a9-dea2d0faf998 was re-scheduled: Binding failed for port 7fffd140-832b-417c-a31f-6d533c144442, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 570.612345] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 570.612345] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.612345] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.612345] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.612345] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.815s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.612618] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.612618] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 570.613815] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.201s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.619269] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f55d0e82-cc9b-4246-9933-ed2bcdf0f9de {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.628276] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d29039-6429-4aa9-871c-8068c5c094f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.641203] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2586d877-e107-4585-b3ec-87f742fe5bef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.672111] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e901c3-1b7b-4123-a796-6338c3667d31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.681575] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2a8c5d4f-6e2e-452e-848c-69e09c991124 could not be found. [ 570.681718] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 570.681897] env[61911]: INFO nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Took 0.07 seconds to destroy the instance on the hypervisor. [ 570.682166] env[61911]: DEBUG oslo.service.loopingcall [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.682667] env[61911]: DEBUG nova.compute.manager [-] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 570.682763] env[61911]: DEBUG nova.network.neutron [-] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 570.697355] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693f561d-24f8-4dc9-8ce3-775aa002dc3f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.706071] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b832d643-48bc-4056-ad10-78d7368adb5d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.742042] env[61911]: DEBUG nova.network.neutron [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.743486] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181492MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 570.743725] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.825351] env[61911]: DEBUG nova.network.neutron [-] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.895609] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d7e988bb-0560-40d1-88b2-d27a39e0d8c0 tempest-ServerDiagnosticsNegativeTest-211312230 tempest-ServerDiagnosticsNegativeTest-211312230-project-member] Lock "02ca2bcc-3a06-4e3d-9985-c158b541cdcc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.165s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.896955] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "02ca2bcc-3a06-4e3d-9985-c158b541cdcc" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 16.606s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.897166] env[61911]: INFO nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 02ca2bcc-3a06-4e3d-9985-c158b541cdcc] During sync_power_state the instance has a pending task (spawning). Skip. [ 570.897346] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "02ca2bcc-3a06-4e3d-9985-c158b541cdcc" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.068727] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.159797] env[61911]: DEBUG nova.compute.manager [req-06d051c8-b7d2-4d8e-bbe5-b7a179d8c883 req-c73261d1-9a25-4423-8b08-bbf6ee71ffe3 service nova] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Received event network-vif-deleted-54603ce5-e295-4258-8877-4387d9f81d28 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 571.209291] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.246139] env[61911]: INFO nova.compute.manager [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] Took 1.13 seconds to deallocate network for instance. [ 571.316774] env[61911]: DEBUG nova.compute.manager [req-b666c9d9-bfb0-4c6e-91a7-78fd80f4cd18 req-aec3f39b-b87f-4e76-a90a-3b630d6fd9e8 service nova] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Received event network-vif-deleted-f419ee8d-147a-45b7-bf04-96850a6effa7 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 571.327703] env[61911]: DEBUG nova.network.neutron [-] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.400955] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 571.428120] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf871f1-392a-4b10-ae54-06a745399282 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.439826] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e81e3a-28b8-42dd-a220-5b1920357617 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.471816] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a856cfb-3e3b-4401-b17b-af9a03dc4059 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.481142] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319a9ae0-5721-4884-b189-d2902483dadd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.496570] env[61911]: DEBUG nova.compute.provider_tree [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.548366] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.576294] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Releasing lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.576294] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 571.576294] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.576294] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f731254c-91ee-4dee-b7e9-d7ff765a083f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.590472] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f083fe8-3192-45e3-8fb6-2aa6a9b6619d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.627623] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 20ca33cb-1bf2-443a-9a35-31dcd01b4266 could not be found. [ 571.627876] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.628289] env[61911]: INFO nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Took 0.05 seconds to destroy the instance on the hypervisor. [ 571.628544] env[61911]: DEBUG oslo.service.loopingcall [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.628806] env[61911]: DEBUG nova.compute.manager [-] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 571.628861] env[61911]: DEBUG nova.network.neutron [-] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.685882] env[61911]: DEBUG nova.network.neutron [-] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.830030] env[61911]: INFO nova.compute.manager [-] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Took 1.15 seconds to deallocate network for instance. [ 571.834602] env[61911]: DEBUG nova.compute.claims [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.834602] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.930296] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.999155] env[61911]: DEBUG nova.scheduler.client.report [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 572.051992] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-82e22142-4083-4716-95a9-dea2d0faf998" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.051992] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 572.051992] env[61911]: DEBUG nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 572.051992] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 572.102122] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.188914] env[61911]: DEBUG nova.network.neutron [-] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.300173] env[61911]: INFO nova.scheduler.client.report [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Deleted allocations for instance 54cd80a2-17fb-446d-b25e-d6d68f8bd538 [ 572.313775] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "c4a4e40a-c3d3-4d27-98a4-452f18c6408d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.313775] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "c4a4e40a-c3d3-4d27-98a4-452f18c6408d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.369513] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "23376481-f0eb-4ccf-b68f-1aca112a4a2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.369908] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "23376481-f0eb-4ccf-b68f-1aca112a4a2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.508178] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.509298] env[61911]: ERROR nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Traceback (most recent call last): [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self.driver.spawn(context, instance, image_meta, [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] vm_ref = self.build_virtual_machine(instance, [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.509298] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] for vif in network_info: [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return self._sync_wrapper(fn, *args, **kwargs) [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self.wait() [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self[:] = self._gt.wait() [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return self._exit_event.wait() [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] result = hub.switch() [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.509743] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return self.greenlet.switch() [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] result = function(*args, **kwargs) [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] return func(*args, **kwargs) [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] raise e [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] nwinfo = self.network_api.allocate_for_instance( [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] created_port_ids = self._update_ports_for_instance( [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] with excutils.save_and_reraise_exception(): [ 572.510237] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] self.force_reraise() [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] raise self.value [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] updated_port = self._update_port( [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] _ensure_no_port_binding_failure(port) [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] raise exception.PortBindingFailed(port_id=port['id']) [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] nova.exception.PortBindingFailed: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. [ 572.510612] env[61911]: ERROR nova.compute.manager [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] [ 572.510928] env[61911]: DEBUG nova.compute.utils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.512136] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.250s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.519832] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Build of instance 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba was re-scheduled: Binding failed for port 916ecd84-a31f-42d2-bfba-6253c4382ec9, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 572.520460] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 572.520772] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Acquiring lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.520994] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Acquired lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.521224] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.603627] env[61911]: DEBUG nova.network.neutron [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.696025] env[61911]: INFO nova.compute.manager [-] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Took 1.07 seconds to deallocate network for instance. [ 572.698462] env[61911]: DEBUG nova.compute.claims [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.698652] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.812935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7dee397c-eb72-4cd0-bc65-3a32c2c62286 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "54cd80a2-17fb-446d-b25e-d6d68f8bd538" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.709s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.814294] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "54cd80a2-17fb-446d-b25e-d6d68f8bd538" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 18.523s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.814499] env[61911]: INFO nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 54cd80a2-17fb-446d-b25e-d6d68f8bd538] During sync_power_state the instance has a pending task (spawning). Skip. [ 572.814676] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "54cd80a2-17fb-446d-b25e-d6d68f8bd538" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.074195] env[61911]: ERROR nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. [ 573.074195] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.074195] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.074195] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.074195] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.074195] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.074195] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.074195] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.074195] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.074195] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 573.074195] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.074195] env[61911]: ERROR nova.compute.manager raise self.value [ 573.074195] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.074195] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.074195] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.074195] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.074856] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.074856] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.074856] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. [ 573.074856] env[61911]: ERROR nova.compute.manager [ 573.077022] env[61911]: Traceback (most recent call last): [ 573.077022] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.077022] env[61911]: listener.cb(fileno) [ 573.077022] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.077022] env[61911]: result = function(*args, **kwargs) [ 573.077022] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.077022] env[61911]: return func(*args, **kwargs) [ 573.077022] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 573.077022] env[61911]: raise e [ 573.077022] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.077022] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 573.077022] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.077022] env[61911]: created_port_ids = self._update_ports_for_instance( [ 573.077022] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.077022] env[61911]: with excutils.save_and_reraise_exception(): [ 573.077022] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.077022] env[61911]: self.force_reraise() [ 573.077022] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.077022] env[61911]: raise self.value [ 573.077022] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.077022] env[61911]: updated_port = self._update_port( [ 573.077022] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.077022] env[61911]: _ensure_no_port_binding_failure(port) [ 573.077022] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.077022] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.077022] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. [ 573.077022] env[61911]: Removing descriptor: 16 [ 573.077817] env[61911]: ERROR nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Traceback (most recent call last): [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] yield resources [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self.driver.spawn(context, instance, image_meta, [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] vm_ref = self.build_virtual_machine(instance, [ 573.077817] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] for vif in network_info: [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return self._sync_wrapper(fn, *args, **kwargs) [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self.wait() [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self[:] = self._gt.wait() [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return self._exit_event.wait() [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.078209] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] result = hub.switch() [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return self.greenlet.switch() [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] result = function(*args, **kwargs) [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return func(*args, **kwargs) [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] raise e [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] nwinfo = self.network_api.allocate_for_instance( [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] created_port_ids = self._update_ports_for_instance( [ 573.078558] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] with excutils.save_and_reraise_exception(): [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self.force_reraise() [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] raise self.value [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] updated_port = self._update_port( [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] _ensure_no_port_binding_failure(port) [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] raise exception.PortBindingFailed(port_id=port['id']) [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] nova.exception.PortBindingFailed: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. [ 573.078910] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] [ 573.079275] env[61911]: INFO nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Terminating instance [ 573.108360] env[61911]: INFO nova.compute.manager [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] Took 1.05 seconds to deallocate network for instance. [ 573.145236] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.318190] env[61911]: DEBUG nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 573.330311] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Acquiring lock "f6d437f9-c341-4f4b-9b68-63711ca0ed42" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.330627] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Lock "f6d437f9-c341-4f4b-9b68-63711ca0ed42" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.429820] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5069df-e8de-4745-816e-6b3bc4862d6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.441215] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f345c81-feff-4c15-a150-904fc3e2ddd3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.477600] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf519e32-192a-48b3-9340-58f5d3433562 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.490630] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedd0b40-a0d9-4805-8f2c-b3abe84c6a13 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.509493] env[61911]: DEBUG nova.compute.provider_tree [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.557248] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.587855] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Acquiring lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.588056] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Acquired lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.588235] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 573.853493] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.016690] env[61911]: DEBUG nova.scheduler.client.report [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 574.066162] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Releasing lock "refresh_cache-2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.066162] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 574.066162] env[61911]: DEBUG nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 574.066162] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.112573] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.117979] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.165637] env[61911]: INFO nova.scheduler.client.report [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleted allocations for instance 82e22142-4083-4716-95a9-dea2d0faf998 [ 574.238262] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.417909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Acquiring lock "95733825-2618-4f6a-b2eb-bdef3a7f60de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.417909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Lock "95733825-2618-4f6a-b2eb-bdef3a7f60de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.525232] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.525659] env[61911]: ERROR nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Traceback (most recent call last): [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self.driver.spawn(context, instance, image_meta, [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] vm_ref = self.build_virtual_machine(instance, [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.525659] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] for vif in network_info: [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return self._sync_wrapper(fn, *args, **kwargs) [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self.wait() [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self[:] = self._gt.wait() [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return self._exit_event.wait() [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] result = hub.switch() [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.526143] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return self.greenlet.switch() [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] result = function(*args, **kwargs) [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] return func(*args, **kwargs) [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] raise e [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] nwinfo = self.network_api.allocate_for_instance( [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] created_port_ids = self._update_ports_for_instance( [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] with excutils.save_and_reraise_exception(): [ 574.526618] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] self.force_reraise() [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] raise self.value [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] updated_port = self._update_port( [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] _ensure_no_port_binding_failure(port) [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] raise exception.PortBindingFailed(port_id=port['id']) [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] nova.exception.PortBindingFailed: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. [ 574.527170] env[61911]: ERROR nova.compute.manager [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] [ 574.528682] env[61911]: DEBUG nova.compute.utils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 574.528682] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.351s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.538590] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Build of instance 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66 was re-scheduled: Binding failed for port 29dbc5e9-60eb-47ed-b07f-c125a3587158, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 574.538590] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 574.538590] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Acquiring lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.538590] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Acquired lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.539254] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.619507] env[61911]: DEBUG nova.network.neutron [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.674321] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9c3726b-a484-4cbe-8220-24e696ad1adb tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "82e22142-4083-4716-95a9-dea2d0faf998" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.468s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.675223] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "82e22142-4083-4716-95a9-dea2d0faf998" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.384s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.676061] env[61911]: INFO nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 82e22142-4083-4716-95a9-dea2d0faf998] During sync_power_state the instance has a pending task (spawning). Skip. [ 574.679777] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "82e22142-4083-4716-95a9-dea2d0faf998" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.740918] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Releasing lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.741629] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 574.741985] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 574.746311] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b801661-2bcc-4edd-84ae-ad3a902ccc50 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.761831] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5797d48b-c4f8-47a5-b11b-a3b2a1428879 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.789188] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ad2e485-4482-4514-845b-782c552d30fe could not be found. [ 574.789542] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.789855] env[61911]: INFO nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Took 0.05 seconds to destroy the instance on the hypervisor. [ 574.791042] env[61911]: DEBUG oslo.service.loopingcall [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.791042] env[61911]: DEBUG nova.compute.manager [-] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 574.791042] env[61911]: DEBUG nova.network.neutron [-] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.833234] env[61911]: DEBUG nova.network.neutron [-] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.070604] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.084902] env[61911]: DEBUG nova.compute.manager [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Received event network-changed-43e0a58a-e1c5-4f2a-a643-52815b9757d1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 575.084902] env[61911]: DEBUG nova.compute.manager [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Refreshing instance network info cache due to event network-changed-43e0a58a-e1c5-4f2a-a643-52815b9757d1. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 575.091418] env[61911]: DEBUG oslo_concurrency.lockutils [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] Acquiring lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.091418] env[61911]: DEBUG oslo_concurrency.lockutils [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] Acquired lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.091418] env[61911]: DEBUG nova.network.neutron [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Refreshing network info cache for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 575.122638] env[61911]: INFO nova.compute.manager [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] Took 1.06 seconds to deallocate network for instance. [ 575.178803] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 575.215242] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.342149] env[61911]: DEBUG nova.network.neutron [-] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.358652] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Acquiring lock "17eec473-7082-4103-baf5-b61dc0d63281" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.358904] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Lock "17eec473-7082-4103-baf5-b61dc0d63281" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.370098] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b09f620-e728-4ab6-831f-c380bc293de7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.380026] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae469de-4421-4674-b4d8-6237183a83e1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.412995] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac14bd9-d147-441d-a33d-97180c322a08 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.421358] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20145da-b0ad-4d91-a9d3-13082e95fea4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.435911] env[61911]: DEBUG nova.compute.provider_tree [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.611451] env[61911]: DEBUG nova.network.neutron [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.709742] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.716836] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Releasing lock "refresh_cache-2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.717109] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 575.717301] env[61911]: DEBUG nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 575.717480] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.735245] env[61911]: DEBUG nova.network.neutron [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.779415] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.844780] env[61911]: INFO nova.compute.manager [-] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Took 1.05 seconds to deallocate network for instance. [ 575.847195] env[61911]: DEBUG nova.compute.claims [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.847375] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.941054] env[61911]: DEBUG nova.scheduler.client.report [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 576.186526] env[61911]: INFO nova.scheduler.client.report [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Deleted allocations for instance 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba [ 576.238128] env[61911]: DEBUG oslo_concurrency.lockutils [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] Releasing lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.238425] env[61911]: DEBUG nova.compute.manager [req-b8e0bee9-26e6-421a-b001-0c6bb9851af9 req-4eb93710-9639-4964-a5aa-7ddf1ca2bf03 service nova] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Received event network-vif-deleted-43e0a58a-e1c5-4f2a-a643-52815b9757d1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 576.287367] env[61911]: DEBUG nova.network.neutron [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.446822] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.447499] env[61911]: ERROR nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Traceback (most recent call last): [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self.driver.spawn(context, instance, image_meta, [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] vm_ref = self.build_virtual_machine(instance, [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.447499] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] for vif in network_info: [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return self._sync_wrapper(fn, *args, **kwargs) [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self.wait() [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self[:] = self._gt.wait() [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return self._exit_event.wait() [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] result = hub.switch() [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.447839] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return self.greenlet.switch() [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] result = function(*args, **kwargs) [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] return func(*args, **kwargs) [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] raise e [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] nwinfo = self.network_api.allocate_for_instance( [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] created_port_ids = self._update_ports_for_instance( [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] with excutils.save_and_reraise_exception(): [ 576.448196] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] self.force_reraise() [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] raise self.value [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] updated_port = self._update_port( [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] _ensure_no_port_binding_failure(port) [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] raise exception.PortBindingFailed(port_id=port['id']) [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] nova.exception.PortBindingFailed: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. [ 576.448531] env[61911]: ERROR nova.compute.manager [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] [ 576.451521] env[61911]: DEBUG nova.compute.utils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.452314] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.927s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.455613] env[61911]: INFO nova.compute.claims [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.455908] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Build of instance 960d5c25-2a54-4eb8-b175-27e56b9894d9 was re-scheduled: Binding failed for port 80c3cf50-bf99-454f-a37a-6ba0bbd9ac1d, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 576.456383] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 576.456610] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.456754] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquired lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.456909] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.698441] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94029672-2f91-4693-a33f-2256184a4537 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362 tempest-FloatingIPsAssociationNegativeTestJSON-1658183362-project-member] Lock "2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.750s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.698441] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 22.406s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.698441] env[61911]: INFO nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba] During sync_power_state the instance has a pending task (spawning). Skip. [ 576.698441] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "2a9c481f-b45a-4539-bbc1-c10d5ee8a7ba" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.791943] env[61911]: INFO nova.compute.manager [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] Took 1.07 seconds to deallocate network for instance. [ 576.940035] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Acquiring lock "91e485bf-4751-4820-8f77-47bd79cc8423" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.940035] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Lock "91e485bf-4751-4820-8f77-47bd79cc8423" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.993134] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.082906] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.202716] env[61911]: DEBUG nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 577.593047] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Releasing lock "refresh_cache-960d5c25-2a54-4eb8-b175-27e56b9894d9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.593047] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 577.593047] env[61911]: DEBUG nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 577.593047] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.628516] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.735965] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.820526] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aae0d2e-4a9b-44a0-bcca-71085b3ed7ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.829836] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482bd6ce-d77c-40f0-a425-6df4b9b3c6c1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.835025] env[61911]: INFO nova.scheduler.client.report [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Deleted allocations for instance 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66 [ 577.872169] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878b6689-d27d-4f93-b05d-fdd8d0652c78 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.881984] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04d343e-4e84-4991-baf6-d36721381110 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.898448] env[61911]: DEBUG nova.compute.provider_tree [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.136932] env[61911]: DEBUG nova.network.neutron [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.347938] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e0ddfb26-4f0b-452e-a262-1585d6c8ba30 tempest-ServerExternalEventsTest-720077728 tempest-ServerExternalEventsTest-720077728-project-member] Lock "2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.111s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.348390] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 24.057s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.348579] env[61911]: INFO nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 2363f012-ab6d-40d3-bd9d-f4ed38e8bc66] During sync_power_state the instance has a pending task (networking). Skip. [ 578.348753] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "2363f012-ab6d-40d3-bd9d-f4ed38e8bc66" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.404123] env[61911]: DEBUG nova.scheduler.client.report [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 578.641132] env[61911]: INFO nova.compute.manager [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] Took 1.05 seconds to deallocate network for instance. [ 578.655094] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Acquiring lock "0a592ff4-3de3-41a2-9f34-a552203b66a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.655333] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Lock "0a592ff4-3de3-41a2-9f34-a552203b66a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.701718] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "bf3d8ee0-3da2-4d18-8fe5-0472118491a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.701924] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "bf3d8ee0-3da2-4d18-8fe5-0472118491a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.850970] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 578.909054] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.909812] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 578.913790] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.169s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.381988] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.419053] env[61911]: DEBUG nova.compute.utils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.423499] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 579.424013] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 579.493444] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "5538f850-908d-4c89-8089-98daf1709bc1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.493683] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "5538f850-908d-4c89-8089-98daf1709bc1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.689262] env[61911]: INFO nova.scheduler.client.report [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Deleted allocations for instance 960d5c25-2a54-4eb8-b175-27e56b9894d9 [ 579.712055] env[61911]: DEBUG nova.policy [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '495b10b81f954e86a4a2735fb5e0c7ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85df730fbab84afeb9ec0bcfb0c03c99', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.924033] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 579.968388] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 2a8c5d4f-6e2e-452e-848c-69e09c991124 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 579.970098] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 20ca33cb-1bf2-443a-9a35-31dcd01b4266 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 579.971124] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 0ad2e485-4482-4514-845b-782c552d30fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 579.971332] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 38fcf50f-04c8-494b-9eb0-f62b25e200b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 580.200533] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11a236c7-d457-49f0-8838-8d2d0cf01966 tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "960d5c25-2a54-4eb8-b175-27e56b9894d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.890s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.205689] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "960d5c25-2a54-4eb8-b175-27e56b9894d9" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 25.914s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.206191] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5fffa36-b2a4-4863-956b-770d8a8121cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.222250] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0596a65f-3f71-4045-9847-57fd43a66bfe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.473750] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 9f1af494-bb6c-4f7e-8218-0413848ceb5c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 580.519610] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Successfully created port: 763246bd-4189-493d-bee9-dad3802b8a88 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.707541] env[61911]: DEBUG nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 580.756181] env[61911]: INFO nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 960d5c25-2a54-4eb8-b175-27e56b9894d9] During the sync_power process the instance has moved from host None to host cpu-1 [ 580.757008] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "960d5c25-2a54-4eb8-b175-27e56b9894d9" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.551s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.933936] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 580.942030] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Acquiring lock "00a4e54a-8198-402f-a1ad-2a233480bda8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.942378] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Lock "00a4e54a-8198-402f-a1ad-2a233480bda8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.967800] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 580.968083] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.968247] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 580.968432] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.971475] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 580.971475] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 580.971475] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 580.971475] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 580.971475] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 580.971789] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 580.971789] env[61911]: DEBUG nova.virt.hardware [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 580.971789] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5b7071-7ec5-4378-8a90-6ce90abcb3fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.976382] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 821d3f6f-e847-4ddb-ac00-ea55af302383 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 580.983730] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7eafde8-3717-4774-a592-1f5d9b762927 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.233599] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.484191] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 2bb4e0d0-8154-45d8-b662-02e5a9c28b21 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 581.644213] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "f21b2aae-9d93-4d48-ba19-9905149eb5ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.644652] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "f21b2aae-9d93-4d48-ba19-9905149eb5ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.985763] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 90e2a5db-918a-4f66-a9c8-ef41dc4b855b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.489776] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8da2407a-c7e5-437a-ab71-57b92d98182f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 582.763015] env[61911]: ERROR nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. [ 582.763015] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.763015] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.763015] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.763015] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.763015] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.763015] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.763015] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.763015] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.763015] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 582.763015] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.763015] env[61911]: ERROR nova.compute.manager raise self.value [ 582.763015] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.763015] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.763015] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.763015] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.763545] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.763545] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.763545] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. [ 582.763545] env[61911]: ERROR nova.compute.manager [ 582.763545] env[61911]: Traceback (most recent call last): [ 582.763545] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.763545] env[61911]: listener.cb(fileno) [ 582.763545] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.763545] env[61911]: result = function(*args, **kwargs) [ 582.763545] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.763545] env[61911]: return func(*args, **kwargs) [ 582.763545] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.763545] env[61911]: raise e [ 582.763545] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.763545] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 582.763545] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.763545] env[61911]: created_port_ids = self._update_ports_for_instance( [ 582.763545] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.763545] env[61911]: with excutils.save_and_reraise_exception(): [ 582.763545] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.763545] env[61911]: self.force_reraise() [ 582.763545] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.763545] env[61911]: raise self.value [ 582.763545] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.763545] env[61911]: updated_port = self._update_port( [ 582.763545] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.763545] env[61911]: _ensure_no_port_binding_failure(port) [ 582.763545] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.763545] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.764431] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. [ 582.764431] env[61911]: Removing descriptor: 19 [ 582.764431] env[61911]: ERROR nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Traceback (most recent call last): [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] yield resources [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self.driver.spawn(context, instance, image_meta, [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.764431] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] vm_ref = self.build_virtual_machine(instance, [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] for vif in network_info: [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return self._sync_wrapper(fn, *args, **kwargs) [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self.wait() [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self[:] = self._gt.wait() [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return self._exit_event.wait() [ 582.764828] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] result = hub.switch() [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return self.greenlet.switch() [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] result = function(*args, **kwargs) [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return func(*args, **kwargs) [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] raise e [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] nwinfo = self.network_api.allocate_for_instance( [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.765240] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] created_port_ids = self._update_ports_for_instance( [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] with excutils.save_and_reraise_exception(): [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self.force_reraise() [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] raise self.value [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] updated_port = self._update_port( [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] _ensure_no_port_binding_failure(port) [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.765800] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] raise exception.PortBindingFailed(port_id=port['id']) [ 582.766276] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] nova.exception.PortBindingFailed: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. [ 582.766276] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] [ 582.766276] env[61911]: INFO nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Terminating instance [ 582.993311] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 583.269137] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Acquiring lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.269676] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Acquired lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.269990] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.496279] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance c4a4e40a-c3d3-4d27-98a4-452f18c6408d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 583.845192] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.999214] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 23376481-f0eb-4ccf-b68f-1aca112a4a2d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.153775] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.352855] env[61911]: DEBUG nova.compute.manager [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Received event network-changed-763246bd-4189-493d-bee9-dad3802b8a88 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 584.353072] env[61911]: DEBUG nova.compute.manager [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Refreshing instance network info cache due to event network-changed-763246bd-4189-493d-bee9-dad3802b8a88. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 584.353233] env[61911]: DEBUG oslo_concurrency.lockutils [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] Acquiring lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.502692] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance f6d437f9-c341-4f4b-9b68-63711ca0ed42 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 584.657208] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Releasing lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.657654] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 584.657882] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.658587] env[61911]: DEBUG oslo_concurrency.lockutils [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] Acquired lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.658765] env[61911]: DEBUG nova.network.neutron [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Refreshing network info cache for port 763246bd-4189-493d-bee9-dad3802b8a88 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 584.661728] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8a860f2-83c4-47c5-89cb-05b61ec807b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.676224] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca840c2-4da2-4016-9028-8b8f08145c82 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.705687] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 38fcf50f-04c8-494b-9eb0-f62b25e200b7 could not be found. [ 584.705938] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.706135] env[61911]: INFO nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Took 0.05 seconds to destroy the instance on the hypervisor. [ 584.706382] env[61911]: DEBUG oslo.service.loopingcall [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.706592] env[61911]: DEBUG nova.compute.manager [-] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 584.706682] env[61911]: DEBUG nova.network.neutron [-] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.728631] env[61911]: DEBUG nova.network.neutron [-] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.918686] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Acquiring lock "5235a412-c833-4f7c-a884-2855e61b8966" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.918866] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Lock "5235a412-c833-4f7c-a884-2855e61b8966" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.009228] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 95733825-2618-4f6a-b2eb-bdef3a7f60de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 585.195344] env[61911]: DEBUG nova.network.neutron [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.231807] env[61911]: DEBUG nova.network.neutron [-] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.356789] env[61911]: DEBUG nova.network.neutron [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.512406] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 17eec473-7082-4103-baf5-b61dc0d63281 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 585.736370] env[61911]: INFO nova.compute.manager [-] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Took 1.03 seconds to deallocate network for instance. [ 585.738482] env[61911]: DEBUG nova.compute.claims [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.738711] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.865674] env[61911]: DEBUG oslo_concurrency.lockutils [req-081f11c9-189c-48d9-9924-89b0c5b8328c req-31872403-e3d7-4070-9810-10133bba553c service nova] Releasing lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.021244] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 91e485bf-4751-4820-8f77-47bd79cc8423 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 586.113154] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Acquiring lock "5002bdba-3519-4caa-8486-c2249bd9abef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.113485] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Lock "5002bdba-3519-4caa-8486-c2249bd9abef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.524225] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 0a592ff4-3de3-41a2-9f34-a552203b66a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.028141] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bf3d8ee0-3da2-4d18-8fe5-0472118491a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.272907] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "61c4a92c-296a-4c0e-aa9a-32b20a837b6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.272907] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "61c4a92c-296a-4c0e-aa9a-32b20a837b6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.333079] env[61911]: DEBUG nova.compute.manager [req-3a64cf96-75ab-43cf-8352-38d1c13b080e req-0bf13a26-5749-48b8-b9eb-9e61888a9491 service nova] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Received event network-vif-deleted-763246bd-4189-493d-bee9-dad3802b8a88 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 587.531584] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5538f850-908d-4c89-8089-98daf1709bc1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 587.531858] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 587.531969] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 587.959065] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Acquiring lock "08e38cae-cf76-4367-ac66-c5ecd25c2763" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.959174] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Lock "08e38cae-cf76-4367-ac66-c5ecd25c2763" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.993462] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932cb831-08f3-428a-8f02-e6dbcf5f2e94 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.006022] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec42d9b-7305-4eb0-b13b-d5d4b66ab7d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.037080] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ab6f00-62bd-4bc8-a571-6473dfdf57dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.048151] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e289cd-b8d8-43ae-b11d-badddfad1bed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.063766] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.084330] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Acquiring lock "c24159b8-8974-499e-8237-26f16e3a198b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.084499] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Lock "c24159b8-8974-499e-8237-26f16e3a198b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.567107] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 589.075848] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 589.076123] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.163s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.076469] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.243s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.079586] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 589.080104] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Getting list of instances from cluster (obj){ [ 589.080104] env[61911]: value = "domain-c8" [ 589.080104] env[61911]: _type = "ClusterComputeResource" [ 589.080104] env[61911]: } {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 589.081375] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7237ed8-816e-4a74-9107-d5eb9e730839 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.092214] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Got total of 0 instances {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 589.684109] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "fe6c9377-f9e6-43d2-afa7-2323bf60a48d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.684381] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "fe6c9377-f9e6-43d2-afa7-2323bf60a48d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.034237] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9902ff2a-04dc-4646-bc14-646c65532325 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.043778] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8077b352-851c-4dd9-9d01-5efc5f7e4d85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.085370] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbc3d63-dc18-4403-b556-703c70f45ae9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.101223] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242d798e-2bf5-4549-b2c4-0a467e15554a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.112665] env[61911]: DEBUG nova.compute.provider_tree [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.619463] env[61911]: DEBUG nova.scheduler.client.report [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 591.128035] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.128035] env[61911]: ERROR nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. [ 591.128035] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Traceback (most recent call last): [ 591.128035] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 591.128035] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self.driver.spawn(context, instance, image_meta, [ 591.128035] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 591.128035] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.128035] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.128035] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] vm_ref = self.build_virtual_machine(instance, [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] for vif in network_info: [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return self._sync_wrapper(fn, *args, **kwargs) [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self.wait() [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self[:] = self._gt.wait() [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return self._exit_event.wait() [ 591.128692] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] result = hub.switch() [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return self.greenlet.switch() [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] result = function(*args, **kwargs) [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] return func(*args, **kwargs) [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] raise e [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] nwinfo = self.network_api.allocate_for_instance( [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.129085] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] created_port_ids = self._update_ports_for_instance( [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] with excutils.save_and_reraise_exception(): [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] self.force_reraise() [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] raise self.value [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] updated_port = self._update_port( [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] _ensure_no_port_binding_failure(port) [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.129606] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] raise exception.PortBindingFailed(port_id=port['id']) [ 591.129997] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] nova.exception.PortBindingFailed: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. [ 591.129997] env[61911]: ERROR nova.compute.manager [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] [ 591.129997] env[61911]: DEBUG nova.compute.utils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.132200] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.202s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.134347] env[61911]: INFO nova.compute.claims [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 591.137333] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Build of instance 2a8c5d4f-6e2e-452e-848c-69e09c991124 was re-scheduled: Binding failed for port 54603ce5-e295-4258-8877-4387d9f81d28, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 591.138157] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 591.138157] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Acquiring lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.138263] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Acquired lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.139158] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.683958] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.836051] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.970655] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "e5f4224f-0002-4b75-9f69-d1ac50d955ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.970948] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "e5f4224f-0002-4b75-9f69-d1ac50d955ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.002547] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.002732] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.038293] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "a09bc118-c3f8-4943-89a1-8a0c59d528c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.038814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "a09bc118-c3f8-4943-89a1-8a0c59d528c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.341583] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Releasing lock "refresh_cache-2a8c5d4f-6e2e-452e-848c-69e09c991124" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.341841] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 592.341979] env[61911]: DEBUG nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 592.342202] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.388983] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.570019] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b838e85-7c3e-477a-9f63-94c59dc12311 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.581128] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538f04fd-24b2-43bc-992d-9e6f29299a70 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.613047] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae08537-a404-46a5-a602-58f5934e5dde {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.623033] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cded36-d1f2-47b1-a175-d5c6e7535982 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.638080] env[61911]: DEBUG nova.compute.provider_tree [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.891527] env[61911]: DEBUG nova.network.neutron [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.143134] env[61911]: DEBUG nova.scheduler.client.report [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 593.396599] env[61911]: INFO nova.compute.manager [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] [instance: 2a8c5d4f-6e2e-452e-848c-69e09c991124] Took 1.05 seconds to deallocate network for instance. [ 593.650923] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.651470] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 593.655949] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.956s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.711219] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Acquiring lock "ade5af6e-0235-407c-85e6-9668f50ccad3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.711219] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Lock "ade5af6e-0235-407c-85e6-9668f50ccad3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.162805] env[61911]: DEBUG nova.compute.utils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 594.162805] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 594.162805] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 594.263611] env[61911]: DEBUG nova.policy [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37599e774937409fa43e14ff921a3f8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c954c090262949fabf2755223c006f1a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.437114] env[61911]: INFO nova.scheduler.client.report [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Deleted allocations for instance 2a8c5d4f-6e2e-452e-848c-69e09c991124 [ 594.655706] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988af79e-2754-469e-b599-673841605025 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.665531] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cb67b1-bab7-4560-84b5-1307377d8a00 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.669989] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 594.706045] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463f21ef-4081-4e16-82f5-65013cbd5565 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.716296] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46383b06-e396-49e7-a9bf-102ca59b9bbe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.732502] env[61911]: DEBUG nova.compute.provider_tree [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.950308] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2bc56830-a28b-4957-b4a6-fd66e673a914 tempest-ServersAdminNegativeTestJSON-1144376803 tempest-ServersAdminNegativeTestJSON-1144376803-project-member] Lock "2a8c5d4f-6e2e-452e-848c-69e09c991124" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.877s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.119060] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Successfully created port: a870bfe2-5f92-4572-aabf-60719b42c8d9 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.238674] env[61911]: DEBUG nova.scheduler.client.report [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 595.452982] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 595.680320] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 595.719118] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 595.719328] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.719485] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 595.719756] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.719929] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 595.722535] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 595.722807] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 595.722977] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 595.723169] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 595.723333] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 595.723503] env[61911]: DEBUG nova.virt.hardware [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 595.724445] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ad2008-08cb-43ff-8029-2e8b38516ff9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.733606] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05596c2d-9fef-4206-8e8f-bcd0e2891848 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.746435] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.746435] env[61911]: ERROR nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. [ 595.746435] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Traceback (most recent call last): [ 595.746435] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 595.746435] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self.driver.spawn(context, instance, image_meta, [ 595.746435] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 595.746435] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.746435] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.746435] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] vm_ref = self.build_virtual_machine(instance, [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] for vif in network_info: [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return self._sync_wrapper(fn, *args, **kwargs) [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self.wait() [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self[:] = self._gt.wait() [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return self._exit_event.wait() [ 595.746735] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] result = hub.switch() [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return self.greenlet.switch() [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] result = function(*args, **kwargs) [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] return func(*args, **kwargs) [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] raise e [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] nwinfo = self.network_api.allocate_for_instance( [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.747109] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] created_port_ids = self._update_ports_for_instance( [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] with excutils.save_and_reraise_exception(): [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] self.force_reraise() [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] raise self.value [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] updated_port = self._update_port( [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] _ensure_no_port_binding_failure(port) [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.747685] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] raise exception.PortBindingFailed(port_id=port['id']) [ 595.748096] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] nova.exception.PortBindingFailed: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. [ 595.748096] env[61911]: ERROR nova.compute.manager [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] [ 595.748096] env[61911]: DEBUG nova.compute.utils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.748096] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.893s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.748375] env[61911]: INFO nova.compute.claims [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.751816] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Build of instance 20ca33cb-1bf2-443a-9a35-31dcd01b4266 was re-scheduled: Binding failed for port f419ee8d-147a-45b7-bf04-96850a6effa7, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 595.752309] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 595.752691] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.752760] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquired lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.752943] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.982061] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.271429] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.353062] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.855873] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Releasing lock "refresh_cache-20ca33cb-1bf2-443a-9a35-31dcd01b4266" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.856196] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 596.856306] env[61911]: DEBUG nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 596.856478] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.913489] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.187509] env[61911]: ERROR nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. [ 597.187509] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.187509] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.187509] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.187509] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.187509] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.187509] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.187509] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.187509] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.187509] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 597.187509] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.187509] env[61911]: ERROR nova.compute.manager raise self.value [ 597.187509] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.187509] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.187509] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.187509] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.188190] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.188190] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.188190] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. [ 597.188190] env[61911]: ERROR nova.compute.manager [ 597.188190] env[61911]: Traceback (most recent call last): [ 597.188190] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.188190] env[61911]: listener.cb(fileno) [ 597.188190] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.188190] env[61911]: result = function(*args, **kwargs) [ 597.188190] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.188190] env[61911]: return func(*args, **kwargs) [ 597.188190] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.188190] env[61911]: raise e [ 597.188190] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.188190] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 597.188190] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.188190] env[61911]: created_port_ids = self._update_ports_for_instance( [ 597.188190] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.188190] env[61911]: with excutils.save_and_reraise_exception(): [ 597.188190] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.188190] env[61911]: self.force_reraise() [ 597.188190] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.188190] env[61911]: raise self.value [ 597.188190] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.188190] env[61911]: updated_port = self._update_port( [ 597.188190] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.188190] env[61911]: _ensure_no_port_binding_failure(port) [ 597.188190] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.188190] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.188994] env[61911]: nova.exception.PortBindingFailed: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. [ 597.188994] env[61911]: Removing descriptor: 19 [ 597.188994] env[61911]: ERROR nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Traceback (most recent call last): [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] yield resources [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self.driver.spawn(context, instance, image_meta, [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.188994] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] vm_ref = self.build_virtual_machine(instance, [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] for vif in network_info: [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return self._sync_wrapper(fn, *args, **kwargs) [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self.wait() [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self[:] = self._gt.wait() [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return self._exit_event.wait() [ 597.189359] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] result = hub.switch() [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return self.greenlet.switch() [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] result = function(*args, **kwargs) [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return func(*args, **kwargs) [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] raise e [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] nwinfo = self.network_api.allocate_for_instance( [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.189755] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] created_port_ids = self._update_ports_for_instance( [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] with excutils.save_and_reraise_exception(): [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self.force_reraise() [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] raise self.value [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] updated_port = self._update_port( [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] _ensure_no_port_binding_failure(port) [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.190144] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] raise exception.PortBindingFailed(port_id=port['id']) [ 597.190473] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] nova.exception.PortBindingFailed: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. [ 597.190473] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] [ 597.190473] env[61911]: INFO nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Terminating instance [ 597.209244] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22df56e0-8b8e-46e5-a36a-3dca6852397b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.220236] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661ddcfc-cd9e-4416-a990-a0ec0b3ca895 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.254264] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73baa18b-5bba-42ee-9937-ab6cefe29ece {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.263328] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc378071-0901-437e-91f8-407d4054b375 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.277797] env[61911]: DEBUG nova.compute.provider_tree [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.413430] env[61911]: DEBUG nova.network.neutron [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.577971] env[61911]: DEBUG nova.compute.manager [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Received event network-changed-a870bfe2-5f92-4572-aabf-60719b42c8d9 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 597.577971] env[61911]: DEBUG nova.compute.manager [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Refreshing instance network info cache due to event network-changed-a870bfe2-5f92-4572-aabf-60719b42c8d9. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 597.577971] env[61911]: DEBUG oslo_concurrency.lockutils [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] Acquiring lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.580761] env[61911]: DEBUG oslo_concurrency.lockutils [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] Acquired lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.581043] env[61911]: DEBUG nova.network.neutron [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Refreshing network info cache for port a870bfe2-5f92-4572-aabf-60719b42c8d9 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 597.694026] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Acquiring lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.780889] env[61911]: DEBUG nova.scheduler.client.report [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 597.916069] env[61911]: INFO nova.compute.manager [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 20ca33cb-1bf2-443a-9a35-31dcd01b4266] Took 1.06 seconds to deallocate network for instance. [ 598.106169] env[61911]: DEBUG nova.network.neutron [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.256963] env[61911]: DEBUG nova.network.neutron [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.288097] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.288681] env[61911]: DEBUG nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 598.291790] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.582s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.293370] env[61911]: INFO nova.compute.claims [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.759777] env[61911]: DEBUG oslo_concurrency.lockutils [req-f9bd7bbe-3eef-45f5-8479-b417bb7d40ff req-8eb4ca28-6518-47ee-b6be-5ebc5483ecf9 service nova] Releasing lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.760286] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Acquired lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.760585] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.798452] env[61911]: DEBUG nova.compute.utils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.799845] env[61911]: DEBUG nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Not allocating networking since 'none' was specified. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 598.944631] env[61911]: INFO nova.scheduler.client.report [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Deleted allocations for instance 20ca33cb-1bf2-443a-9a35-31dcd01b4266 [ 599.287189] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.296553] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Acquiring lock "7f777a69-d8ce-4a2f-83f5-085fad6f9809" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.297023] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Lock "7f777a69-d8ce-4a2f-83f5-085fad6f9809" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.303148] env[61911]: DEBUG nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 599.410699] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.457595] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16047c8-ee51-45bd-947b-40a93947f298 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "20ca33cb-1bf2-443a-9a35-31dcd01b4266" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.756s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.702851] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c46a98-4eff-4b9b-a006-d8d3d439b84f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.710720] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6de4540-82c1-442d-b414-c3a470dd8a2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.744594] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43e7366-6639-4e1a-8453-4f4965ec3329 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.752579] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b11847-e3d2-4af5-83d0-e34b6f245a85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.772306] env[61911]: DEBUG nova.compute.provider_tree [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.915905] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Releasing lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.919779] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 599.919779] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.919779] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-410b811b-22b3-41dc-84c5-e058fd78ce8a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.930732] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a197f3-6cad-441a-9d8a-b2931421639d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.956631] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f1af494-bb6c-4f7e-8218-0413848ceb5c could not be found. [ 599.957339] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.957654] env[61911]: INFO nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.957957] env[61911]: DEBUG oslo.service.loopingcall [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.958262] env[61911]: DEBUG nova.compute.manager [-] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 599.958451] env[61911]: DEBUG nova.network.neutron [-] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.962246] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 599.980581] env[61911]: DEBUG nova.network.neutron [-] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.120889] env[61911]: DEBUG nova.compute.manager [req-589f5ec5-c590-4e2f-9325-47a16490d9d0 req-255f48d6-dace-4945-9373-56083f768db5 service nova] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Received event network-vif-deleted-a870bfe2-5f92-4572-aabf-60719b42c8d9 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 600.276128] env[61911]: DEBUG nova.scheduler.client.report [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 600.319061] env[61911]: DEBUG nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 600.347248] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 600.347497] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.347651] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 600.347888] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.348051] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 600.348210] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 600.348416] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 600.348573] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 600.348740] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 600.348901] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 600.349081] env[61911]: DEBUG nova.virt.hardware [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 600.349964] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca8e38a-7429-4454-ae13-9d87131015a3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.366262] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1dacc7a-1f57-44d9-a982-d24b201ff6d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.383766] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 600.396719] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.397061] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36e7d58d-d7ff-43fa-aee7-154169f7bbd5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.412370] env[61911]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 600.412549] env[61911]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61911) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 600.412901] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 600.413114] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Creating folder: Project (5ddd3bf25e824087972ca32173555acf). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.413351] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4428d3a1-cad1-481f-8347-a7a6ac579093 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.422682] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Created folder: Project (5ddd3bf25e824087972ca32173555acf) in parent group-v269521. [ 600.423026] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Creating folder: Instances. Parent ref: group-v269530. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.423299] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-944ef17c-6605-4d7c-b164-340900c40d2f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.434391] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Created folder: Instances in parent group-v269530. [ 600.434391] env[61911]: DEBUG oslo.service.loopingcall [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.434391] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 600.434391] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73305db0-ab9e-4390-aecc-8e5f060c4b9a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.449511] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 600.449511] env[61911]: value = "task-1250834" [ 600.449511] env[61911]: _type = "Task" [ 600.449511] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.458144] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250834, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.483465] env[61911]: DEBUG nova.network.neutron [-] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.489273] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.781975] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.782204] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 600.788162] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.941s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.966058] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250834, 'name': CreateVM_Task, 'duration_secs': 0.278022} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.966338] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 600.971017] env[61911]: DEBUG oslo_vmware.service [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d02451-e51a-4722-947a-2ce9bc1c643c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.973149] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.973315] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.973931] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 600.974208] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bf012ca-f3c9-406f-9580-61fac873f69a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.979837] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 600.979837] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52971af5-90e4-04ec-385b-071111cbfa0d" [ 600.979837] env[61911]: _type = "Task" [ 600.979837] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.989648] env[61911]: INFO nova.compute.manager [-] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Took 1.03 seconds to deallocate network for instance. [ 600.989648] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52971af5-90e4-04ec-385b-071111cbfa0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.993726] env[61911]: DEBUG nova.compute.claims [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.993944] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.296886] env[61911]: DEBUG nova.compute.utils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.302129] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 601.302320] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 601.409218] env[61911]: DEBUG nova.policy [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '652cc07d037b49449fc5fb24e002649d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8d6ca77ff7d4d52a8d9c6f81333908a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.491998] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.492348] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 601.492528] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.492674] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.493113] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 601.495522] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8788c8d2-39c8-4af8-b4e8-041f8bb75e98 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.505876] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 601.506072] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 601.506850] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d05cc2a-f78f-4bea-a541-829dc6263a1b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.516329] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b981b1bc-0233-4f4c-a977-ff9357255f45 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.523154] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 601.523154] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526e547b-71d0-0346-1168-883032cb20c4" [ 601.523154] env[61911]: _type = "Task" [ 601.523154] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.529920] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526e547b-71d0-0346-1168-883032cb20c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.799192] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38bc7a8-b119-4d40-901d-f6c13c605f0e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.806162] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 601.811899] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15610ae-6113-4953-bb19-6cc15a462925 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.850683] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef87c480-0d65-4751-b609-e248e92c3ceb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.859824] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28ec85d-7bcb-45d0-853e-e76ae73cf08d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.873864] env[61911]: DEBUG nova.compute.provider_tree [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.033557] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Successfully created port: 618f77ed-5e90-47b5-b69d-c7cb033c7f31 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.035614] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Preparing fetch location {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 602.035859] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Creating directory with path [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 602.036111] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-229e3990-e443-4fc3-b669-72740481ef2e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.049067] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Created directory with path [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 602.049067] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Fetch image to [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 602.049067] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Downloading image file data 0dfe2ff1-43fd-4529-93f1-daaccc0711cf to [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk on the data store datastore1 {{(pid=61911) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 602.049067] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822d9832-9217-468c-9ef6-2d5b7c5e9cb9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.056189] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a786a0f-c5a7-436d-9452-6ac1ee938d14 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.065772] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5659e243-0358-429e-9a4c-11dcaac8e81d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.096977] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a697028b-4b04-4c3a-b06a-30d02e0cfe5b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.103358] env[61911]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-0fe6d72c-1be2-4eb6-92a2-0f2ddb225f80 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.129632] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Downloading image file data 0dfe2ff1-43fd-4529-93f1-daaccc0711cf to the data store datastore1 {{(pid=61911) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 602.200603] env[61911]: DEBUG oslo_vmware.rw_handles [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61911) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 602.379862] env[61911]: DEBUG nova.scheduler.client.report [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 602.825021] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 602.854328] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:54:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='178700505',id=26,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1576601171',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 602.855546] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.855859] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 602.856188] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.856449] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 602.856723] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 602.857122] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 602.857418] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 602.858093] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 602.858272] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 602.860017] env[61911]: DEBUG nova.virt.hardware [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 602.860017] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2676b441-e884-4b18-9f1e-550f1ad02ec1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.868211] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4c38a1-80d9-4630-b71c-5312c9dc3afb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.885152] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.097s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.885980] env[61911]: ERROR nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Traceback (most recent call last): [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self.driver.spawn(context, instance, image_meta, [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] vm_ref = self.build_virtual_machine(instance, [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.885980] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] for vif in network_info: [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return self._sync_wrapper(fn, *args, **kwargs) [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self.wait() [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self[:] = self._gt.wait() [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return self._exit_event.wait() [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] result = hub.switch() [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.886412] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return self.greenlet.switch() [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] result = function(*args, **kwargs) [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] return func(*args, **kwargs) [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] raise e [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] nwinfo = self.network_api.allocate_for_instance( [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] created_port_ids = self._update_ports_for_instance( [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] with excutils.save_and_reraise_exception(): [ 602.886825] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] self.force_reraise() [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] raise self.value [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] updated_port = self._update_port( [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] _ensure_no_port_binding_failure(port) [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] raise exception.PortBindingFailed(port_id=port['id']) [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] nova.exception.PortBindingFailed: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. [ 602.887256] env[61911]: ERROR nova.compute.manager [instance: 0ad2e485-4482-4514-845b-782c552d30fe] [ 602.890366] env[61911]: DEBUG nova.compute.utils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.890366] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.154s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.894017] env[61911]: INFO nova.compute.claims [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.903014] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Build of instance 0ad2e485-4482-4514-845b-782c552d30fe was re-scheduled: Binding failed for port 43e0a58a-e1c5-4f2a-a643-52815b9757d1, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 602.903014] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 602.903014] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Acquiring lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.903014] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Acquired lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.903387] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.919221] env[61911]: DEBUG oslo_vmware.rw_handles [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Completed reading data from the image iterator. {{(pid=61911) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 602.919592] env[61911]: DEBUG oslo_vmware.rw_handles [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 602.978687] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Downloaded image file data 0dfe2ff1-43fd-4529-93f1-daaccc0711cf to vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk on the data store datastore1 {{(pid=61911) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 602.981221] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Caching image {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 602.981221] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Copying Virtual Disk [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk to [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 602.983747] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f12c2184-88da-46dc-9d7d-4c759d4864ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.991151] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 602.991151] env[61911]: value = "task-1250835" [ 602.991151] env[61911]: _type = "Task" [ 602.991151] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.002432] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.424760] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.486993] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.505644] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250835, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.991117] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Releasing lock "refresh_cache-0ad2e485-4482-4514-845b-782c552d30fe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.991326] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 603.991493] env[61911]: DEBUG nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.991657] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.010541] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64175} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.010803] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Copied Virtual Disk [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk to [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 604.010984] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleting the datastore file [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 604.011243] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1576bfd-9db2-49b1-ba24-087910172630 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.019539] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 604.019539] env[61911]: value = "task-1250836" [ 604.019539] env[61911]: _type = "Task" [ 604.019539] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.025757] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.034973] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.042150] env[61911]: ERROR nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. [ 604.042150] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.042150] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 604.042150] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.042150] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.042150] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.042150] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.042150] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.042150] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.042150] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 604.042150] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.042150] env[61911]: ERROR nova.compute.manager raise self.value [ 604.042150] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.042150] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.042150] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.042150] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.042817] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.042817] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.042817] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. [ 604.042817] env[61911]: ERROR nova.compute.manager [ 604.042817] env[61911]: Traceback (most recent call last): [ 604.042817] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.042817] env[61911]: listener.cb(fileno) [ 604.042817] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.042817] env[61911]: result = function(*args, **kwargs) [ 604.042817] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.042817] env[61911]: return func(*args, **kwargs) [ 604.042817] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 604.042817] env[61911]: raise e [ 604.042817] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 604.042817] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 604.042817] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.042817] env[61911]: created_port_ids = self._update_ports_for_instance( [ 604.042817] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.042817] env[61911]: with excutils.save_and_reraise_exception(): [ 604.042817] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.042817] env[61911]: self.force_reraise() [ 604.042817] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.042817] env[61911]: raise self.value [ 604.042817] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.042817] env[61911]: updated_port = self._update_port( [ 604.042817] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.042817] env[61911]: _ensure_no_port_binding_failure(port) [ 604.042817] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.042817] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.043971] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. [ 604.043971] env[61911]: Removing descriptor: 19 [ 604.043971] env[61911]: ERROR nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Traceback (most recent call last): [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] yield resources [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self.driver.spawn(context, instance, image_meta, [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.043971] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] vm_ref = self.build_virtual_machine(instance, [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] for vif in network_info: [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return self._sync_wrapper(fn, *args, **kwargs) [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self.wait() [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self[:] = self._gt.wait() [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return self._exit_event.wait() [ 604.044705] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] result = hub.switch() [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return self.greenlet.switch() [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] result = function(*args, **kwargs) [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return func(*args, **kwargs) [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] raise e [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] nwinfo = self.network_api.allocate_for_instance( [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.045102] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] created_port_ids = self._update_ports_for_instance( [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] with excutils.save_and_reraise_exception(): [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self.force_reraise() [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] raise self.value [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] updated_port = self._update_port( [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] _ensure_no_port_binding_failure(port) [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.045572] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] raise exception.PortBindingFailed(port_id=port['id']) [ 604.046341] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] nova.exception.PortBindingFailed: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. [ 604.046341] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] [ 604.046341] env[61911]: INFO nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Terminating instance [ 604.098537] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Acquiring lock "ea27990b-d194-4b81-b18b-37804c22ceb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.098767] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Lock "ea27990b-d194-4b81-b18b-37804c22ceb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.355637] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cca6d56-dbca-4aef-986b-8b3a34bf6afc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.364253] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08d1f2c-cd13-496b-9d12-66b2e57ea34f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.394780] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc12972c-18ae-41f3-9984-faf519e45b1a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.401797] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b616bebd-3d86-4601-a109-d404e3476f4a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.414832] env[61911]: DEBUG nova.compute.provider_tree [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.444854] env[61911]: DEBUG nova.compute.manager [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Received event network-changed-618f77ed-5e90-47b5-b69d-c7cb033c7f31 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 604.445119] env[61911]: DEBUG nova.compute.manager [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Refreshing instance network info cache due to event network-changed-618f77ed-5e90-47b5-b69d-c7cb033c7f31. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 604.445271] env[61911]: DEBUG oslo_concurrency.lockutils [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] Acquiring lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.445411] env[61911]: DEBUG oslo_concurrency.lockutils [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] Acquired lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.445567] env[61911]: DEBUG nova.network.neutron [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Refreshing network info cache for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 604.531999] env[61911]: DEBUG nova.network.neutron [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.533117] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.05468} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.533539] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 604.533742] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Moving file from [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262/0dfe2ff1-43fd-4529-93f1-daaccc0711cf to [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf. {{(pid=61911) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 604.534015] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-4f55858c-64a5-471b-9f1c-4543671bc00a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.540718] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 604.540718] env[61911]: value = "task-1250837" [ 604.540718] env[61911]: _type = "Task" [ 604.540718] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.548987] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250837, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.552542] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Acquiring lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.918384] env[61911]: DEBUG nova.scheduler.client.report [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 604.975059] env[61911]: DEBUG nova.network.neutron [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.034574] env[61911]: INFO nova.compute.manager [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] [instance: 0ad2e485-4482-4514-845b-782c552d30fe] Took 1.04 seconds to deallocate network for instance. [ 605.051289] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250837, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023893} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.051569] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] File moved {{(pid=61911) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 605.051766] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Cleaning up location [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 605.051929] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleting the datastore file [datastore1] vmware_temp/b2c045bf-4f09-4ff9-b3a2-53988bda6262 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 605.052193] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4e25254-9f40-4773-a2ee-7e151c71cd5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.060749] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 605.060749] env[61911]: value = "task-1250838" [ 605.060749] env[61911]: _type = "Task" [ 605.060749] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.070166] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250838, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.139071] env[61911]: DEBUG nova.network.neutron [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.423045] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.423608] env[61911]: DEBUG nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 605.426761] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.044s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.428356] env[61911]: INFO nova.compute.claims [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.572502] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250838, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025344} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.572919] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 605.575394] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-def09644-acda-461b-9b38-cdc220aabbf0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.578834] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 605.578834] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52025b37-ce26-ad6e-743d-75eff39dce05" [ 605.578834] env[61911]: _type = "Task" [ 605.578834] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.591784] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52025b37-ce26-ad6e-743d-75eff39dce05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.641876] env[61911]: DEBUG oslo_concurrency.lockutils [req-59cac815-dfe5-43b7-bdb1-dfcce3f9afad req-866fdcee-f992-4c12-82a9-fdc19ccbaeed service nova] Releasing lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.642644] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Acquired lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.642755] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.933779] env[61911]: DEBUG nova.compute.utils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.937411] env[61911]: DEBUG nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Not allocating networking since 'none' was specified. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 606.070635] env[61911]: INFO nova.scheduler.client.report [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Deleted allocations for instance 0ad2e485-4482-4514-845b-782c552d30fe [ 606.091572] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52025b37-ce26-ad6e-743d-75eff39dce05, 'name': SearchDatastore_Task, 'duration_secs': 0.008766} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.092844] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.092844] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 606.092844] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79a33d83-cc64-4f24-8763-8b9565194758 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.099307] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 606.099307] env[61911]: value = "task-1250839" [ 606.099307] env[61911]: _type = "Task" [ 606.099307] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.110282] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.182496] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.441083] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.441583] env[61911]: DEBUG nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 606.581436] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b319742-42ff-4553-8e3d-fd3f154b57f7 tempest-TenantUsagesTestJSON-1017866737 tempest-TenantUsagesTestJSON-1017866737-project-member] Lock "0ad2e485-4482-4514-845b-782c552d30fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.179s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.608251] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450609} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.610554] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 606.610784] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 606.611170] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f2f2f86-6856-48a2-bc7b-08dbd72ccc7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.617173] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 606.617173] env[61911]: value = "task-1250840" [ 606.617173] env[61911]: _type = "Task" [ 606.617173] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.626750] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.877196] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1418c68b-d982-47df-92a4-3bfb2c9d02c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.885247] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1cca93d-f2cd-4cea-82e8-4373daa88130 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.922973] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f06fa2c-f6ed-4f43-8bcb-23ff0ca0a443 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.928855] env[61911]: DEBUG nova.compute.manager [req-7d04e205-58b4-4c2f-b749-1d2dfc33a0cd req-fdb8475d-0613-4451-aa1a-b1bacf033826 service nova] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Received event network-vif-deleted-618f77ed-5e90-47b5-b69d-c7cb033c7f31 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 606.932623] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e85c84-2eab-439e-80c5-4ff7479e5758 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.947135] env[61911]: DEBUG nova.compute.provider_tree [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.950066] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Releasing lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.950462] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 606.950653] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.953751] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ead01bd4-0f4a-46de-8a08-c06348136224 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.962512] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc1082e-14d5-4a89-a214-21eee3d1290a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.986037] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2bb4e0d0-8154-45d8-b662-02e5a9c28b21 could not be found. [ 606.986037] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.986037] env[61911]: INFO nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Took 0.04 seconds to destroy the instance on the hypervisor. [ 606.986361] env[61911]: DEBUG oslo.service.loopingcall [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.986802] env[61911]: DEBUG nova.compute.manager [-] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 606.986896] env[61911]: DEBUG nova.network.neutron [-] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.015453] env[61911]: DEBUG nova.network.neutron [-] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.083994] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 607.126715] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061019} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.127777] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 607.128710] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed492daa-e2e7-4fa6-8752-df9210407005 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.156285] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 607.156955] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-256891fb-8e3c-441a-bfaa-11f38f212e4a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.179053] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 607.179053] env[61911]: value = "task-1250841" [ 607.179053] env[61911]: _type = "Task" [ 607.179053] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.186675] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250841, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.450157] env[61911]: DEBUG nova.scheduler.client.report [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 607.457499] env[61911]: DEBUG nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 607.492752] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 607.492994] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.493174] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 607.493358] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.493546] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 607.493713] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 607.493926] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 607.495059] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 607.495059] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 607.495059] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 607.495059] env[61911]: DEBUG nova.virt.hardware [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 607.495841] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b557cb38-eb57-4a5a-b05c-f75c0389fb7b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.504617] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6e0314-1043-43a1-98ba-c1765b04eb78 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.518742] env[61911]: DEBUG nova.network.neutron [-] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.520903] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 607.525755] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Creating folder: Project (acb58c586d514661a70788a325963b03). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 607.526359] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-025a7a95-39c4-4665-b792-ceeb68845439 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.537074] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Created folder: Project (acb58c586d514661a70788a325963b03) in parent group-v269521. [ 607.537307] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Creating folder: Instances. Parent ref: group-v269533. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 607.537550] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9985dad4-78ba-4d68-bd3c-383b2fd0dfa0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.545525] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Created folder: Instances in parent group-v269533. [ 607.545748] env[61911]: DEBUG oslo.service.loopingcall [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.545926] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 607.546136] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22d9abb4-6a84-4f4f-950c-a991cc3de4bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.562996] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 607.562996] env[61911]: value = "task-1250844" [ 607.562996] env[61911]: _type = "Task" [ 607.562996] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.571249] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250844, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.612956] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.692591] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250841, 'name': ReconfigVM_Task, 'duration_secs': 0.287762} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.692591] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 607.692591] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ac8106e-eb35-4e1d-a707-aa6fe3062608 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.699237] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 607.699237] env[61911]: value = "task-1250845" [ 607.699237] env[61911]: _type = "Task" [ 607.699237] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.706953] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250845, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.956579] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.957827] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 607.961742] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.728s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.967707] env[61911]: INFO nova.compute.claims [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.027835] env[61911]: INFO nova.compute.manager [-] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Took 1.04 seconds to deallocate network for instance. [ 608.031413] env[61911]: DEBUG nova.compute.claims [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.031750] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.076771] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250844, 'name': CreateVM_Task, 'duration_secs': 0.491228} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.077795] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 608.078688] env[61911]: DEBUG oslo_vmware.service [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9085439-100b-4f56-aba2-b62b815fc891 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.085825] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.085825] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.085825] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 608.086044] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff75af8a-fafe-4805-ac57-204a7040ceb4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.093886] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 608.093886] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524feff9-1ef3-de3d-2016-0d33b056a4d7" [ 608.093886] env[61911]: _type = "Task" [ 608.093886] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.099469] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524feff9-1ef3-de3d-2016-0d33b056a4d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.215179] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250845, 'name': Rename_Task, 'duration_secs': 0.258944} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.216221] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 608.216984] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1183b657-ab60-4e69-8445-f6e2a9528e17 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.227405] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 608.227405] env[61911]: value = "task-1250846" [ 608.227405] env[61911]: _type = "Task" [ 608.227405] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.237277] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.463227] env[61911]: DEBUG nova.compute.utils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.464844] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 608.465887] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 608.549734] env[61911]: DEBUG nova.policy [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd8ce53e94d09461eabcd67ab834178ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80aef6024398403c9273ea3b0e19d839', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.610981] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.611224] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 608.611477] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.611623] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.611795] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 608.612448] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07c0da34-eced-4f42-bb09-0b1bfd8db8c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.621885] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 608.622257] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 608.623318] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9212418a-085f-4334-99f4-e424878fef3b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.631160] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dec1794-79b7-4262-a24d-f20285f14e7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.638596] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 608.638596] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524cf31d-0bdf-a8b7-f0c0-4bb5e7da004f" [ 608.638596] env[61911]: _type = "Task" [ 608.638596] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.649824] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524cf31d-0bdf-a8b7-f0c0-4bb5e7da004f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.742385] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250846, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.969773] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 609.057519] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Successfully created port: 0b3c0ab0-0be3-4503-b5ad-159f20c14904 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.155283] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Preparing fetch location {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 609.155624] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Creating directory with path [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 609.155920] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34ac8d30-64e5-49d4-94c0-ebfbb3d8b331 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.183155] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Created directory with path [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 609.183155] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Fetch image to [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 609.183155] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Downloading image file data 0dfe2ff1-43fd-4529-93f1-daaccc0711cf to [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk on the data store datastore2 {{(pid=61911) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 609.183155] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caed71b7-e94c-4932-aeb1-78d74d5fea6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.196846] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd415881-5b94-4f3e-992f-6865d78ce791 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.214813] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881671d2-cafd-48ae-bc1d-6cbf707b02d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.260442] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c36403f-d52d-4c8c-87ba-afd06e8a7dc8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.269822] env[61911]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-2ac3c0d8-c63c-4544-86e2-7fee9a1bd67d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.272562] env[61911]: DEBUG oslo_vmware.api [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250846, 'name': PowerOnVM_Task, 'duration_secs': 0.896554} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.272870] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 609.273086] env[61911]: INFO nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Took 8.96 seconds to spawn the instance on the hypervisor. [ 609.273417] env[61911]: DEBUG nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 609.274432] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dec0499-6df3-49b6-b94f-d481c2283c2e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.295192] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Downloading image file data 0dfe2ff1-43fd-4529-93f1-daaccc0711cf to the data store datastore2 {{(pid=61911) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 609.363280] env[61911]: DEBUG oslo_vmware.rw_handles [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61911) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 609.631094] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c510c8f5-bef8-4a4d-a5cc-a4fddb3348c6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.639424] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4f7fbc-118f-4f37-aaea-d4ae18ae803d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.675084] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488214fc-c5c9-41e1-8f9e-aa2a86291062 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.686133] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb48fad-e6f7-41f3-b02b-2f23d62a180b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.700487] env[61911]: DEBUG nova.compute.provider_tree [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.799129] env[61911]: INFO nova.compute.manager [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Took 35.97 seconds to build instance. [ 609.984805] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 610.008017] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 610.008017] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.008017] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 610.008228] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.008228] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 610.008228] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 610.008492] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 610.008845] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 610.009182] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 610.009522] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 610.010566] env[61911]: DEBUG nova.virt.hardware [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 610.011104] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3301d096-a7e4-4b25-b377-36c9bfa8e739 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.022152] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ad21ce-3e25-4cad-9c0e-ca8733d00a0f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.094281] env[61911]: DEBUG oslo_vmware.rw_handles [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Completed reading data from the image iterator. {{(pid=61911) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 610.094520] env[61911]: DEBUG oslo_vmware.rw_handles [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 610.206117] env[61911]: DEBUG nova.scheduler.client.report [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 610.236484] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Downloaded image file data 0dfe2ff1-43fd-4529-93f1-daaccc0711cf to vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk on the data store datastore2 {{(pid=61911) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 610.238546] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Caching image {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 610.238812] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Copying Virtual Disk [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk to [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 610.239088] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28a5dbb9-063b-4e61-babc-731a7956665b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.247148] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 610.247148] env[61911]: value = "task-1250847" [ 610.247148] env[61911]: _type = "Task" [ 610.247148] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 610.255077] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.305069] env[61911]: DEBUG oslo_concurrency.lockutils [None req-36ad1f05-3d5a-430e-beb6-e4a50913c36b tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "821d3f6f-e847-4ddb-ac00-ea55af302383" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.681s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.711543] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.750s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.712107] env[61911]: DEBUG nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 610.714752] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.976s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.757790] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250847, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.776594] env[61911]: ERROR nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. [ 610.776594] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.776594] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 610.776594] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.776594] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.776594] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.776594] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.776594] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.776594] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.776594] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 610.776594] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.776594] env[61911]: ERROR nova.compute.manager raise self.value [ 610.776594] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.776594] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.776594] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.776594] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.777201] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.777201] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.777201] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. [ 610.777201] env[61911]: ERROR nova.compute.manager [ 610.777201] env[61911]: Traceback (most recent call last): [ 610.777201] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.777201] env[61911]: listener.cb(fileno) [ 610.777201] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.777201] env[61911]: result = function(*args, **kwargs) [ 610.777201] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.777201] env[61911]: return func(*args, **kwargs) [ 610.777201] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 610.777201] env[61911]: raise e [ 610.777201] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 610.777201] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 610.777201] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.777201] env[61911]: created_port_ids = self._update_ports_for_instance( [ 610.777201] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.777201] env[61911]: with excutils.save_and_reraise_exception(): [ 610.777201] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.777201] env[61911]: self.force_reraise() [ 610.777201] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.777201] env[61911]: raise self.value [ 610.777201] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.777201] env[61911]: updated_port = self._update_port( [ 610.777201] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.777201] env[61911]: _ensure_no_port_binding_failure(port) [ 610.777201] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.777201] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.778204] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. [ 610.778204] env[61911]: Removing descriptor: 19 [ 610.778610] env[61911]: ERROR nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Traceback (most recent call last): [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] yield resources [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self.driver.spawn(context, instance, image_meta, [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] vm_ref = self.build_virtual_machine(instance, [ 610.778610] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] for vif in network_info: [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return self._sync_wrapper(fn, *args, **kwargs) [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self.wait() [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self[:] = self._gt.wait() [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return self._exit_event.wait() [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.778943] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] result = hub.switch() [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return self.greenlet.switch() [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] result = function(*args, **kwargs) [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return func(*args, **kwargs) [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] raise e [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] nwinfo = self.network_api.allocate_for_instance( [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] created_port_ids = self._update_ports_for_instance( [ 610.779396] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] with excutils.save_and_reraise_exception(): [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self.force_reraise() [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] raise self.value [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] updated_port = self._update_port( [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] _ensure_no_port_binding_failure(port) [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] raise exception.PortBindingFailed(port_id=port['id']) [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] nova.exception.PortBindingFailed: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. [ 610.779818] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] [ 610.780197] env[61911]: INFO nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Terminating instance [ 610.807327] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 610.865305] env[61911]: DEBUG nova.compute.manager [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Received event network-changed-0b3c0ab0-0be3-4503-b5ad-159f20c14904 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 610.865305] env[61911]: DEBUG nova.compute.manager [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Refreshing instance network info cache due to event network-changed-0b3c0ab0-0be3-4503-b5ad-159f20c14904. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 610.865305] env[61911]: DEBUG oslo_concurrency.lockutils [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] Acquiring lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.865524] env[61911]: DEBUG oslo_concurrency.lockutils [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] Acquired lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.865587] env[61911]: DEBUG nova.network.neutron [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Refreshing network info cache for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.219796] env[61911]: DEBUG nova.compute.utils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.225397] env[61911]: DEBUG nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Not allocating networking since 'none' was specified. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 611.263306] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673865} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.263306] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Copied Virtual Disk [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk to [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 611.263306] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Deleting the datastore file [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/tmp-sparse.vmdk {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 611.263541] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ef05f24-314e-4821-b95e-c12a1807f9a3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.270813] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 611.270813] env[61911]: value = "task-1250848" [ 611.270813] env[61911]: _type = "Task" [ 611.270813] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.278379] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250848, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.293813] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Acquiring lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.329463] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.414448] env[61911]: DEBUG nova.network.neutron [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.623442] env[61911]: INFO nova.compute.manager [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Rebuilding instance [ 611.671294] env[61911]: DEBUG nova.compute.manager [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 611.672210] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee94a20f-5c65-4bb5-9c0b-4bacf76003c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.676762] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46075f4e-36cf-4b1d-b01d-a1d1cb1e73a7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.687441] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090ffc5d-e2d3-42fc-aea4-43dc26d02db5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.716580] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64127f93-3679-4719-bffd-8e3f7308c7ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.724253] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c96f93-b3c3-4a57-9030-d63e0960d322 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.729857] env[61911]: DEBUG nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 611.742185] env[61911]: DEBUG nova.compute.provider_tree [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.767216] env[61911]: DEBUG nova.network.neutron [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.780069] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250848, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087124} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.780360] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 611.780622] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Moving file from [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366/0dfe2ff1-43fd-4529-93f1-daaccc0711cf to [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf. {{(pid=61911) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 611.780920] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-bfb1300b-8937-413f-83c6-b4b324595652 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.789096] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 611.789096] env[61911]: value = "task-1250849" [ 611.789096] env[61911]: _type = "Task" [ 611.789096] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.799705] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250849, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.244842] env[61911]: DEBUG nova.scheduler.client.report [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 612.269795] env[61911]: DEBUG oslo_concurrency.lockutils [req-0154fe27-3ee1-4079-9af7-cb33993b45de req-bcf115b3-7bb8-4888-ba00-4bd7678807ff service nova] Releasing lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.272987] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Acquired lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.272987] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.304703] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250849, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.032913} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.305378] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] File moved {{(pid=61911) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 612.305746] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Cleaning up location [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 612.306090] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Deleting the datastore file [datastore2] vmware_temp/16656f2d-9a79-457a-b14a-d16234d40366 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 612.306488] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ed3b721-fc68-4035-87d1-aff6902e35ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.313021] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 612.313021] env[61911]: value = "task-1250850" [ 612.313021] env[61911]: _type = "Task" [ 612.313021] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.321411] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.689606] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 612.689959] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-595a260f-2eab-44cb-8cfb-8b5fb402c4be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.697318] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 612.697318] env[61911]: value = "task-1250851" [ 612.697318] env[61911]: _type = "Task" [ 612.697318] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.706705] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.724491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.724491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.742632] env[61911]: DEBUG nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 612.752043] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.752043] env[61911]: ERROR nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. [ 612.752043] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Traceback (most recent call last): [ 612.752043] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 612.752043] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self.driver.spawn(context, instance, image_meta, [ 612.752043] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 612.752043] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.752043] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.752043] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] vm_ref = self.build_virtual_machine(instance, [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] for vif in network_info: [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return self._sync_wrapper(fn, *args, **kwargs) [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self.wait() [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self[:] = self._gt.wait() [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return self._exit_event.wait() [ 612.752404] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] result = hub.switch() [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return self.greenlet.switch() [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] result = function(*args, **kwargs) [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] return func(*args, **kwargs) [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] raise e [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] nwinfo = self.network_api.allocate_for_instance( [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.752773] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] created_port_ids = self._update_ports_for_instance( [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] with excutils.save_and_reraise_exception(): [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] self.force_reraise() [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] raise self.value [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] updated_port = self._update_port( [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] _ensure_no_port_binding_failure(port) [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.753152] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] raise exception.PortBindingFailed(port_id=port['id']) [ 612.753497] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] nova.exception.PortBindingFailed: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. [ 612.753497] env[61911]: ERROR nova.compute.manager [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] [ 612.753497] env[61911]: DEBUG nova.compute.utils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.754186] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.772s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.755662] env[61911]: INFO nova.compute.claims [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.758323] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Build of instance 38fcf50f-04c8-494b-9eb0-f62b25e200b7 was re-scheduled: Binding failed for port 763246bd-4189-493d-bee9-dad3802b8a88, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 612.758771] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 612.759028] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Acquiring lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.759176] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Acquired lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.759332] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.770620] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 612.770816] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.771051] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 612.771301] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.771461] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 612.771610] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 612.771823] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 612.772143] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 612.772411] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 612.772642] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 612.772909] env[61911]: DEBUG nova.virt.hardware [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 612.773895] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419a6a78-a7fc-4c99-af15-6c18ddc9d607 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.789575] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac95e3c-9e9b-4e70-a35c-114a439f0584 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.809437] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 612.816598] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Creating folder: Project (8fd28d9338384e0ebf46a9f2a0290d51). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.817432] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.819444] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8dbfb920-4b01-4f7c-85cb-09d6979801fb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.832121] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023411} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.833495] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 612.834281] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Created folder: Project (8fd28d9338384e0ebf46a9f2a0290d51) in parent group-v269521. [ 612.834491] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Creating folder: Instances. Parent ref: group-v269536. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.834744] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39304a31-d395-49f6-9f48-760ef122b0e5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.836954] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8dd447eb-8957-47bd-965b-2181796c15ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.841861] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 612.841861] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e32a3b-6b02-7b27-d4cd-90652f0bbdfd" [ 612.841861] env[61911]: _type = "Task" [ 612.841861] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.846860] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Created folder: Instances in parent group-v269536. [ 612.847206] env[61911]: DEBUG oslo.service.loopingcall [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.847974] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 612.848270] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3668db00-96b1-4793-92da-3373b93905f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.865386] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e32a3b-6b02-7b27-d4cd-90652f0bbdfd, 'name': SearchDatastore_Task, 'duration_secs': 0.00897} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.866052] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.866385] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 90e2a5db-918a-4f66-a9c8-ef41dc4b855b/90e2a5db-918a-4f66-a9c8-ef41dc4b855b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 612.866677] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c54bc82-c424-4c19-95a3-4842f63b51bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.869697] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 612.869697] env[61911]: value = "task-1250854" [ 612.869697] env[61911]: _type = "Task" [ 612.869697] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.877439] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250854, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.878890] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 612.878890] env[61911]: value = "task-1250855" [ 612.878890] env[61911]: _type = "Task" [ 612.878890] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.888879] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.927606] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.209071] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250851, 'name': PowerOffVM_Task, 'duration_secs': 0.128934} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.209431] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 613.209704] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.210521] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321da9ee-47d1-46ab-b7a5-82c33a1fbc46 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.218239] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 613.218525] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-613ba0a7-2b22-442f-b5e6-63c853d4ae20 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.243534] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 613.243746] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 613.243922] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleting the datastore file [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 613.244206] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5294060-d29f-4f0c-8d19-aee7ab36bfbd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.250598] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 613.250598] env[61911]: value = "task-1250857" [ 613.250598] env[61911]: _type = "Task" [ 613.250598] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.259730] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250857, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.267143] env[61911]: DEBUG nova.compute.manager [req-54cff66b-645d-4bf6-901d-283a9ef47508 req-16944416-fd93-4590-b77b-6bb2369ffcd5 service nova] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Received event network-vif-deleted-0b3c0ab0-0be3-4503-b5ad-159f20c14904 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 613.289051] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.384482] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250854, 'name': CreateVM_Task, 'duration_secs': 0.344399} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.387949] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 613.388467] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.388621] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.388937] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 613.391290] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6268b6b-3ebb-40fa-8456-2181d0959956 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.395634] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250855, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.398943] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 613.398943] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d4c09c-867e-0509-dd2f-92b5939e3bb9" [ 613.398943] env[61911]: _type = "Task" [ 613.398943] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.405729] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d4c09c-867e-0509-dd2f-92b5939e3bb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.435366] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Releasing lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.435366] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 613.435366] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.435366] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9196e995-2d32-4b4d-9d4c-85d9a288ff7c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.439332] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.447241] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb18db9-a18f-4c1d-a53e-bd56026542ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.472868] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8da2407a-c7e5-437a-ab71-57b92d98182f could not be found. [ 613.473191] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.473289] env[61911]: INFO nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.473529] env[61911]: DEBUG oslo.service.loopingcall [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.473750] env[61911]: DEBUG nova.compute.manager [-] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 613.473848] env[61911]: DEBUG nova.network.neutron [-] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.489710] env[61911]: DEBUG nova.network.neutron [-] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.761243] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250857, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204518} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.761506] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 613.761708] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 613.761851] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.900112] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Acquiring lock "fe46e0e7-c1aa-4bf5-af01-cdb04031f945" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.900229] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Lock "fe46e0e7-c1aa-4bf5-af01-cdb04031f945" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.907443] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53831} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.908166] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 90e2a5db-918a-4f66-a9c8-ef41dc4b855b/90e2a5db-918a-4f66-a9c8-ef41dc4b855b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 613.908526] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 613.908768] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da91ddcb-d76f-4ca6-8482-db23508ecbfe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.914618] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d4c09c-867e-0509-dd2f-92b5939e3bb9, 'name': SearchDatastore_Task, 'duration_secs': 0.011306} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.915321] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.915602] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.915869] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.916056] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.916269] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.916748] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64152947-cc2a-4549-bfac-6c5baedb8f03 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.922419] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 613.922419] env[61911]: value = "task-1250858" [ 613.922419] env[61911]: _type = "Task" [ 613.922419] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.929491] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.929846] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 613.934275] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cb1edc2-02f3-46be-a174-c1a398e06149 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.936727] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250858, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.943117] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 613.943117] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52db20c4-d7f7-2746-aa4a-4ac12bb960c8" [ 613.943117] env[61911]: _type = "Task" [ 613.943117] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.947882] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Releasing lock "refresh_cache-38fcf50f-04c8-494b-9eb0-f62b25e200b7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.947882] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 613.948041] env[61911]: DEBUG nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 613.948320] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.955534] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52db20c4-d7f7-2746-aa4a-4ac12bb960c8, 'name': SearchDatastore_Task, 'duration_secs': 0.008873} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.959672] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a4bdcc8-2a04-45da-a20e-97954c1fcaf5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.964505] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 613.964505] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c9c3fc-f1e8-7c8c-6685-442c4f6432da" [ 613.964505] env[61911]: _type = "Task" [ 613.964505] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.969189] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.976041] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c9c3fc-f1e8-7c8c-6685-442c4f6432da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.992275] env[61911]: DEBUG nova.network.neutron [-] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.269682] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3a5d12-fc90-4ac1-baa2-22a4a62fda6d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.277184] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d5b6a1-56b6-4d94-b10f-cfea144a9eb6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.307013] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1e23dd-4af9-47b1-8ee9-f84685be0c4a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.314450] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8281450d-867e-4ccf-9e6a-d26e50f68057 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.328602] env[61911]: DEBUG nova.compute.provider_tree [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.433202] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260969} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.433532] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 614.434317] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbc6583-5b3f-4c3f-8cae-01169d67e6ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.454800] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 90e2a5db-918a-4f66-a9c8-ef41dc4b855b/90e2a5db-918a-4f66-a9c8-ef41dc4b855b.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 614.455183] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7363d750-8ebb-4d57-a9af-543ea120012e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.472688] env[61911]: DEBUG nova.network.neutron [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.478939] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c9c3fc-f1e8-7c8c-6685-442c4f6432da, 'name': SearchDatastore_Task, 'duration_secs': 0.010081} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.480087] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.480339] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 614.480674] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 614.480674] env[61911]: value = "task-1250859" [ 614.480674] env[61911]: _type = "Task" [ 614.480674] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.480879] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22d940e8-a785-4a71-b257-c9bb54d25e0e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.493970] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250859, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.494263] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 614.494263] env[61911]: value = "task-1250860" [ 614.494263] env[61911]: _type = "Task" [ 614.494263] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.498187] env[61911]: INFO nova.compute.manager [-] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Took 1.02 seconds to deallocate network for instance. [ 614.500464] env[61911]: DEBUG nova.compute.claims [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.500464] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.503569] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.799598] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 614.799977] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.800153] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 614.800338] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.800557] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 614.800911] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 614.801231] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 614.801330] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 614.801502] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 614.801668] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 614.801841] env[61911]: DEBUG nova.virt.hardware [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 614.802729] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b8f878-f709-457d-a861-4a2ff04678ed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.812237] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42e51cf-0706-480a-adf2-9641656f9a90 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.828405] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 614.835776] env[61911]: DEBUG oslo.service.loopingcall [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.836949] env[61911]: DEBUG nova.scheduler.client.report [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 614.840709] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 614.845016] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b9857b6-96e0-4be0-bec6-ef86e7a880e7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.866329] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 614.866329] env[61911]: value = "task-1250861" [ 614.866329] env[61911]: _type = "Task" [ 614.866329] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.875163] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250861, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.977044] env[61911]: INFO nova.compute.manager [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] [instance: 38fcf50f-04c8-494b-9eb0-f62b25e200b7] Took 1.03 seconds to deallocate network for instance. [ 614.993743] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250859, 'name': ReconfigVM_Task, 'duration_secs': 0.287031} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.994042] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 90e2a5db-918a-4f66-a9c8-ef41dc4b855b/90e2a5db-918a-4f66-a9c8-ef41dc4b855b.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 614.994638] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0548ee69-bcea-409e-9235-6529a76d51e9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.006096] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250860, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436172} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.007378] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 615.007480] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 615.007890] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 615.007890] env[61911]: value = "task-1250862" [ 615.007890] env[61911]: _type = "Task" [ 615.007890] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.010602] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38abce10-b8d1-499f-8e0b-41a97f611c31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.018192] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250862, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.019466] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 615.019466] env[61911]: value = "task-1250863" [ 615.019466] env[61911]: _type = "Task" [ 615.019466] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.027918] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250863, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.343163] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.343221] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 615.347762] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.858s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.348856] env[61911]: INFO nova.compute.claims [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.375468] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250861, 'name': CreateVM_Task, 'duration_secs': 0.322463} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.375659] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 615.376119] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.376275] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.376585] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 615.376824] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41fe4de2-be19-4642-8379-c7797b1dc1a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.382366] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 615.382366] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52256ae6-0397-99ef-a978-57c4805169ba" [ 615.382366] env[61911]: _type = "Task" [ 615.382366] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.390767] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52256ae6-0397-99ef-a978-57c4805169ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.521405] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250862, 'name': Rename_Task, 'duration_secs': 0.143367} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.524697] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 615.524789] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a22b6a8d-272e-46cc-9d08-0f91ca146ac5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.531221] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250863, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078125} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.532386] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 615.532718] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 615.532718] env[61911]: value = "task-1250864" [ 615.532718] env[61911]: _type = "Task" [ 615.532718] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.533391] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c4b361-0a12-4779-b278-8240c16ffecf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.543591] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250864, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.561189] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 615.561653] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e81ab496-309f-4c17-9118-add29142b07e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.590951] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 615.590951] env[61911]: value = "task-1250865" [ 615.590951] env[61911]: _type = "Task" [ 615.590951] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.599323] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250865, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.853390] env[61911]: DEBUG nova.compute.utils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.857877] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 615.857877] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.892395] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52256ae6-0397-99ef-a978-57c4805169ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009426} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.892586] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.892827] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 615.893554] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.893554] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.893554] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 615.893733] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92529d3b-b7fd-48d7-bda6-473a5c27fde5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.901427] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 615.902394] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 615.903190] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b84352b-fbcf-481a-a426-e1d6dec7e964 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.908556] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 615.908556] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e005cd-f003-13c5-bc55-30bed190929b" [ 615.908556] env[61911]: _type = "Task" [ 615.908556] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.917465] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e005cd-f003-13c5-bc55-30bed190929b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.921022] env[61911]: DEBUG nova.policy [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c57bc181d3f746d39b8beb5e8aae3839', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4806147e49484ae2887767632c518e6d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.005136] env[61911]: INFO nova.scheduler.client.report [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Deleted allocations for instance 38fcf50f-04c8-494b-9eb0-f62b25e200b7 [ 616.045377] env[61911]: DEBUG oslo_vmware.api [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250864, 'name': PowerOnVM_Task, 'duration_secs': 0.423626} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.045377] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 616.045615] env[61911]: INFO nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Took 8.59 seconds to spawn the instance on the hypervisor. [ 616.045798] env[61911]: DEBUG nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 616.046620] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec31b91-d70b-460b-943f-c17fa9f8581b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.102346] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250865, 'name': ReconfigVM_Task, 'duration_secs': 0.252011} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.103967] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Reconfigured VM instance instance-00000011 to attach disk [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 616.103967] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28721701-784e-40ae-a287-b2dce78d053e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.112200] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 616.112200] env[61911]: value = "task-1250866" [ 616.112200] env[61911]: _type = "Task" [ 616.112200] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.123302] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250866, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.299594] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Successfully created port: 2371d3f8-8347-4e79-abd6-4d5882113fb0 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.361022] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 616.422062] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e005cd-f003-13c5-bc55-30bed190929b, 'name': SearchDatastore_Task, 'duration_secs': 0.008482} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.422994] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fb3f2f0-5d1b-43ca-bcde-5ca42934b6f4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.433019] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 616.433019] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e1d82c-11a5-46e4-beeb-286eddd385d5" [ 616.433019] env[61911]: _type = "Task" [ 616.433019] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.440148] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e1d82c-11a5-46e4-beeb-286eddd385d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.512730] env[61911]: DEBUG oslo_concurrency.lockutils [None req-802fd215-9cbe-43c2-be67-a1760307cc38 tempest-ImagesOneServerNegativeTestJSON-740876305 tempest-ImagesOneServerNegativeTestJSON-740876305-project-member] Lock "38fcf50f-04c8-494b-9eb0-f62b25e200b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.789s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.569236] env[61911]: INFO nova.compute.manager [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Took 38.85 seconds to build instance. [ 616.621171] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250866, 'name': Rename_Task, 'duration_secs': 0.139933} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.621449] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.621686] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df61bfc0-c158-45ce-be46-ec08eb714f37 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.628363] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 616.628363] env[61911]: value = "task-1250867" [ 616.628363] env[61911]: _type = "Task" [ 616.628363] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.636261] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.755745] env[61911]: DEBUG nova.compute.manager [None req-3628d34e-8932-4ba1-9638-2687b303786e tempest-ServerDiagnosticsV248Test-1715709720 tempest-ServerDiagnosticsV248Test-1715709720-project-admin] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 616.759955] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156f8a06-be29-434c-a27f-8fe1a32b12dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.769442] env[61911]: INFO nova.compute.manager [None req-3628d34e-8932-4ba1-9638-2687b303786e tempest-ServerDiagnosticsV248Test-1715709720 tempest-ServerDiagnosticsV248Test-1715709720-project-admin] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Retrieving diagnostics [ 616.770521] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92d80ae-0ed3-44a1-a533-34fefd7ec338 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.845505] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5fe595-d7b1-4a77-91ff-b3dffb491f25 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.854025] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fbc913-a306-4632-a391-3afbc4558ba6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.888825] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d59e1b3-896d-4063-a84e-99ffb0c2fb47 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.896541] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4184f3d-441e-409c-a3ed-8811a7e66dce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.910644] env[61911]: DEBUG nova.compute.provider_tree [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.942673] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e1d82c-11a5-46e4-beeb-286eddd385d5, 'name': SearchDatastore_Task, 'duration_secs': 0.008415} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.942673] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.942673] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 616.944814] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-887ccacb-e06b-417c-810e-a11ad617cde7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.950600] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 616.950600] env[61911]: value = "task-1250868" [ 616.950600] env[61911]: _type = "Task" [ 616.950600] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.959967] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250868, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.014908] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 617.071158] env[61911]: DEBUG oslo_concurrency.lockutils [None req-413b1b39-0d55-49ab-98b4-9af5b828957d tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.202s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.143767] env[61911]: DEBUG oslo_vmware.api [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250867, 'name': PowerOnVM_Task, 'duration_secs': 0.462721} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.144408] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 617.144792] env[61911]: INFO nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Took 4.40 seconds to spawn the instance on the hypervisor. [ 617.145174] env[61911]: DEBUG nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 617.146407] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8b3c36-fd96-4e37-bf98-7b5e7c84aeba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.396588] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 617.418025] env[61911]: DEBUG nova.scheduler.client.report [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 617.440012] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 617.440291] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.440429] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 617.440605] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.440795] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 617.440943] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 617.441160] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 617.441309] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 617.441467] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 617.441623] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 617.441964] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 617.443112] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917bb4d6-2a76-47c7-98f5-af9848e031d1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.459975] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c4db35-8d08-4c1d-8974-a6f5dca33af7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.469015] env[61911]: DEBUG nova.compute.manager [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Received event network-changed-2371d3f8-8347-4e79-abd6-4d5882113fb0 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 617.469225] env[61911]: DEBUG nova.compute.manager [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Refreshing instance network info cache due to event network-changed-2371d3f8-8347-4e79-abd6-4d5882113fb0. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 617.469442] env[61911]: DEBUG oslo_concurrency.lockutils [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] Acquiring lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.469586] env[61911]: DEBUG oslo_concurrency.lockutils [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] Acquired lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.469786] env[61911]: DEBUG nova.network.neutron [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Refreshing network info cache for port 2371d3f8-8347-4e79-abd6-4d5882113fb0 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.483584] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250868, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.530281] env[61911]: DEBUG nova.network.neutron [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.545248] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.568475] env[61911]: ERROR nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. [ 617.568475] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.568475] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.568475] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.568475] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.568475] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.568475] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.568475] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.568475] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.568475] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 617.568475] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.568475] env[61911]: ERROR nova.compute.manager raise self.value [ 617.568475] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.568475] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.568475] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.568475] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.569327] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.569327] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.569327] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. [ 617.569327] env[61911]: ERROR nova.compute.manager [ 617.569327] env[61911]: Traceback (most recent call last): [ 617.569327] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.569327] env[61911]: listener.cb(fileno) [ 617.569327] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.569327] env[61911]: result = function(*args, **kwargs) [ 617.569327] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.569327] env[61911]: return func(*args, **kwargs) [ 617.569327] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 617.569327] env[61911]: raise e [ 617.569327] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.569327] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 617.569327] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.569327] env[61911]: created_port_ids = self._update_ports_for_instance( [ 617.569327] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.569327] env[61911]: with excutils.save_and_reraise_exception(): [ 617.569327] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.569327] env[61911]: self.force_reraise() [ 617.569327] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.569327] env[61911]: raise self.value [ 617.569327] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.569327] env[61911]: updated_port = self._update_port( [ 617.569327] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.569327] env[61911]: _ensure_no_port_binding_failure(port) [ 617.569327] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.569327] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.570867] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. [ 617.570867] env[61911]: Removing descriptor: 19 [ 617.570867] env[61911]: ERROR nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Traceback (most recent call last): [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] yield resources [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self.driver.spawn(context, instance, image_meta, [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.570867] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] vm_ref = self.build_virtual_machine(instance, [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] for vif in network_info: [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return self._sync_wrapper(fn, *args, **kwargs) [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self.wait() [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self[:] = self._gt.wait() [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return self._exit_event.wait() [ 617.571381] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] result = hub.switch() [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return self.greenlet.switch() [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] result = function(*args, **kwargs) [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return func(*args, **kwargs) [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] raise e [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] nwinfo = self.network_api.allocate_for_instance( [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.572325] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] created_port_ids = self._update_ports_for_instance( [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] with excutils.save_and_reraise_exception(): [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self.force_reraise() [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] raise self.value [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] updated_port = self._update_port( [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] _ensure_no_port_binding_failure(port) [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.573167] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] raise exception.PortBindingFailed(port_id=port['id']) [ 617.573779] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] nova.exception.PortBindingFailed: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. [ 617.573779] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] [ 617.573779] env[61911]: INFO nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Terminating instance [ 617.574615] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 617.667911] env[61911]: INFO nova.compute.manager [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Took 36.45 seconds to build instance. [ 617.817173] env[61911]: DEBUG nova.network.neutron [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.920720] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.922317] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 617.924284] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.930s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.964874] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250868, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605786} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.965658] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 617.965906] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 617.967022] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e55f9ed-10d2-454e-8a40-5c96343bc622 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.973419] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 617.973419] env[61911]: value = "task-1250869" [ 617.973419] env[61911]: _type = "Task" [ 617.973419] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.981228] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250869, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.074648] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.095529] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.169293] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b664890a-b3bf-4e6a-b0ad-1f53729ddf4d tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.315s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.324022] env[61911]: DEBUG oslo_concurrency.lockutils [req-8737723a-d375-4516-9e7f-e22bd126d9db req-ce87af12-a7f4-4f55-b137-16653ae98c4d service nova] Releasing lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.324022] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.324022] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 618.429950] env[61911]: DEBUG nova.compute.utils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 618.434437] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 618.434625] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 618.492250] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250869, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063143} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.492250] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 618.492250] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6a23fb-92ac-4cf2-b8b5-0117b8ce5000 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.514020] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 618.514945] env[61911]: DEBUG nova.policy [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c57bc181d3f746d39b8beb5e8aae3839', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4806147e49484ae2887767632c518e6d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.518834] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79ca0584-78b6-4c5b-baad-d2cc4261ed57 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.538555] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 618.538555] env[61911]: value = "task-1250870" [ 618.538555] env[61911]: _type = "Task" [ 618.538555] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.548447] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250870, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.675187] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 618.840949] env[61911]: INFO nova.compute.manager [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Rebuilding instance [ 618.850214] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Successfully created port: 0567fcb8-b74b-405a-b78f-4d9413d5d5ae {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.853602] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.907581] env[61911]: DEBUG nova.compute.manager [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 618.908459] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b14c41-12dc-4ceb-844d-0dab346a68d8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.937379] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 618.941763] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ce6776-b465-4816-88c3-aa5702134296 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.949941] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60391ab-4968-4232-bbb0-56e5fb7550cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.982727] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c32623-9ba0-4b38-9263-a74e2c8619d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.987047] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.992139] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0c8f5e-ba65-433d-8af3-feeb8de655f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.007341] env[61911]: DEBUG nova.compute.provider_tree [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.048395] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250870, 'name': ReconfigVM_Task, 'duration_secs': 0.308818} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.048686] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 619.049598] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc6c9292-54e7-42c5-acaa-f71154d48729 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.056315] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 619.056315] env[61911]: value = "task-1250871" [ 619.056315] env[61911]: _type = "Task" [ 619.056315] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.064622] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250871, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.203570] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.490029] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.490479] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 619.490855] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.490995] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f161640-b42d-467d-b4f3-cfd61c9eded8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.503423] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8572eb6c-9ec2-425f-a23d-e5bea616a8b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.517575] env[61911]: DEBUG nova.scheduler.client.report [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 619.525194] env[61911]: DEBUG nova.compute.manager [req-6074a161-e6fd-4a39-bf45-8e17da566dc5 req-753c3f64-b125-4163-b202-37972f9d0b44 service nova] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Received event network-vif-deleted-2371d3f8-8347-4e79-abd6-4d5882113fb0 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 619.543288] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c4a4e40a-c3d3-4d27-98a4-452f18c6408d could not be found. [ 619.543288] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.543288] env[61911]: INFO nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 619.543288] env[61911]: DEBUG oslo.service.loopingcall [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.543288] env[61911]: DEBUG nova.compute.manager [-] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 619.543288] env[61911]: DEBUG nova.network.neutron [-] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 619.565421] env[61911]: DEBUG nova.network.neutron [-] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.572676] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250871, 'name': Rename_Task, 'duration_secs': 0.131} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.573166] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 619.573885] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97ea7fed-6038-4d2e-81a0-49d620e563f5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.580254] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 619.580254] env[61911]: value = "task-1250872" [ 619.580254] env[61911]: _type = "Task" [ 619.580254] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.589018] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250872, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.668686] env[61911]: DEBUG nova.compute.manager [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Received event network-changed-0567fcb8-b74b-405a-b78f-4d9413d5d5ae {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 619.668883] env[61911]: DEBUG nova.compute.manager [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Refreshing instance network info cache due to event network-changed-0567fcb8-b74b-405a-b78f-4d9413d5d5ae. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 619.669115] env[61911]: DEBUG oslo_concurrency.lockutils [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] Acquiring lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.669274] env[61911]: DEBUG oslo_concurrency.lockutils [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] Acquired lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.669459] env[61911]: DEBUG nova.network.neutron [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Refreshing network info cache for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 619.889276] env[61911]: ERROR nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. [ 619.889276] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 619.889276] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.889276] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 619.889276] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.889276] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 619.889276] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.889276] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 619.889276] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.889276] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 619.889276] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.889276] env[61911]: ERROR nova.compute.manager raise self.value [ 619.889276] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.889276] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 619.889276] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.889276] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 619.889958] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.889958] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 619.889958] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. [ 619.889958] env[61911]: ERROR nova.compute.manager [ 619.889958] env[61911]: Traceback (most recent call last): [ 619.889958] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 619.889958] env[61911]: listener.cb(fileno) [ 619.889958] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.889958] env[61911]: result = function(*args, **kwargs) [ 619.889958] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.889958] env[61911]: return func(*args, **kwargs) [ 619.889958] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 619.889958] env[61911]: raise e [ 619.889958] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.889958] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 619.889958] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.889958] env[61911]: created_port_ids = self._update_ports_for_instance( [ 619.889958] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.889958] env[61911]: with excutils.save_and_reraise_exception(): [ 619.889958] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.889958] env[61911]: self.force_reraise() [ 619.889958] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.889958] env[61911]: raise self.value [ 619.889958] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.889958] env[61911]: updated_port = self._update_port( [ 619.889958] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.889958] env[61911]: _ensure_no_port_binding_failure(port) [ 619.889958] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.889958] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 619.890796] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. [ 619.890796] env[61911]: Removing descriptor: 16 [ 619.925297] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 619.925297] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e912206-fd0a-4e39-946a-0472c19d5de7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.932026] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 619.932026] env[61911]: value = "task-1250873" [ 619.932026] env[61911]: _type = "Task" [ 619.932026] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.942243] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250873, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.952978] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 619.985186] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 619.985533] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.985742] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 619.985980] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.986200] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 619.986395] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 619.986655] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 619.986939] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 619.987544] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 619.987544] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 619.987544] env[61911]: DEBUG nova.virt.hardware [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 619.988493] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3378e6b-bf16-45aa-9600-d22b9464dc7b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.998017] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64035489-7a25-4b13-833e-f8c3c2af7f42 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.012831] env[61911]: ERROR nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Traceback (most recent call last): [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] yield resources [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self.driver.spawn(context, instance, image_meta, [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] vm_ref = self.build_virtual_machine(instance, [ 620.012831] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] for vif in network_info: [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] return self._sync_wrapper(fn, *args, **kwargs) [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self.wait() [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self[:] = self._gt.wait() [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] return self._exit_event.wait() [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 620.013239] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] current.throw(*self._exc) [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] result = function(*args, **kwargs) [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] return func(*args, **kwargs) [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] raise e [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] nwinfo = self.network_api.allocate_for_instance( [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] created_port_ids = self._update_ports_for_instance( [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] with excutils.save_and_reraise_exception(): [ 620.013637] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self.force_reraise() [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] raise self.value [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] updated_port = self._update_port( [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] _ensure_no_port_binding_failure(port) [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] raise exception.PortBindingFailed(port_id=port['id']) [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] nova.exception.PortBindingFailed: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. [ 620.014149] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] [ 620.014149] env[61911]: INFO nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Terminating instance [ 620.027374] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.103s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.027997] env[61911]: ERROR nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Traceback (most recent call last): [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self.driver.spawn(context, instance, image_meta, [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] vm_ref = self.build_virtual_machine(instance, [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.027997] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] for vif in network_info: [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return self._sync_wrapper(fn, *args, **kwargs) [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self.wait() [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self[:] = self._gt.wait() [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return self._exit_event.wait() [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] result = hub.switch() [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.028483] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return self.greenlet.switch() [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] result = function(*args, **kwargs) [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] return func(*args, **kwargs) [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] raise e [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] nwinfo = self.network_api.allocate_for_instance( [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] created_port_ids = self._update_ports_for_instance( [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] with excutils.save_and_reraise_exception(): [ 620.028901] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] self.force_reraise() [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] raise self.value [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] updated_port = self._update_port( [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] _ensure_no_port_binding_failure(port) [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] raise exception.PortBindingFailed(port_id=port['id']) [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] nova.exception.PortBindingFailed: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. [ 620.029384] env[61911]: ERROR nova.compute.manager [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] [ 620.029779] env[61911]: DEBUG nova.compute.utils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.031286] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.419s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.033024] env[61911]: INFO nova.compute.claims [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.036297] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Build of instance 9f1af494-bb6c-4f7e-8218-0413848ceb5c was re-scheduled: Binding failed for port a870bfe2-5f92-4572-aabf-60719b42c8d9, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 620.036766] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 620.036995] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Acquiring lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.037159] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Acquired lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.037321] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.074492] env[61911]: DEBUG nova.network.neutron [-] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.090146] env[61911]: DEBUG oslo_vmware.api [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250872, 'name': PowerOnVM_Task, 'duration_secs': 0.459751} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.090446] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 620.090705] env[61911]: DEBUG nova.compute.manager [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 620.091507] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5fb3ab-b61b-49b3-9334-46060b49e642 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.191403] env[61911]: DEBUG nova.network.neutron [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.281750] env[61911]: DEBUG nova.network.neutron [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.440684] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250873, 'name': PowerOffVM_Task, 'duration_secs': 0.193855} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.440955] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 620.441209] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.441950] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fd20d9-36cb-49fd-a19a-60bed7dec46d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.448148] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 620.448364] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a208a228-81ab-4de0-ac61-d3c8d16600e5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.469986] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 620.470216] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 620.470426] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Deleting the datastore file [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 620.470696] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-667e38b3-afa7-4d30-8118-0e67fc0a041a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.477035] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 620.477035] env[61911]: value = "task-1250875" [ 620.477035] env[61911]: _type = "Task" [ 620.477035] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.484297] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.523302] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.562110] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.578156] env[61911]: INFO nova.compute.manager [-] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Took 1.04 seconds to deallocate network for instance. [ 620.580603] env[61911]: DEBUG nova.compute.claims [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 620.580802] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.607826] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.635262] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.786453] env[61911]: DEBUG oslo_concurrency.lockutils [req-6d70edf1-84b9-488c-9e82-a2d50d262a09 req-e6b7116f-6cef-43ae-9069-168a169e77c9 service nova] Releasing lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.786453] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.786453] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.986534] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101325} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.986796] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 620.987049] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 620.987282] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.138596] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Releasing lock "refresh_cache-9f1af494-bb6c-4f7e-8218-0413848ceb5c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.138900] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 621.139130] env[61911]: DEBUG nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 621.139333] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.166341] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.313021] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.447718] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.499052] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13db15d-994c-45e6-af0e-488369a93802 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.506735] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cfcaa5-0ef9-4667-bf5f-135c8eb4cdda {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.538095] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf57b343-f2c7-4d4c-a03a-5c71d5d69931 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.546102] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6510bb5c-c26f-4a50-89f1-74ef01b5e855 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.566365] env[61911]: DEBUG nova.compute.provider_tree [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.666866] env[61911]: DEBUG nova.network.neutron [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.742535] env[61911]: DEBUG nova.compute.manager [req-cf85b414-c4fa-4d92-8230-b72245ec4f2c req-150c3c96-e1cc-4daf-befe-548bb1f42b32 service nova] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Received event network-vif-deleted-0567fcb8-b74b-405a-b78f-4d9413d5d5ae {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 621.953742] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.954193] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 621.954383] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 621.955015] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f047da6e-d423-4708-a490-dff0a85da42d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.964664] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdec196c-b021-4541-b3ad-1f4cc7c8bf33 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.987401] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 23376481-f0eb-4ccf-b68f-1aca112a4a2d could not be found. [ 621.987673] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.987797] env[61911]: INFO nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 621.988044] env[61911]: DEBUG oslo.service.loopingcall [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.988268] env[61911]: DEBUG nova.compute.manager [-] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 621.988360] env[61911]: DEBUG nova.network.neutron [-] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.008974] env[61911]: DEBUG nova.network.neutron [-] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.024099] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 622.024389] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.024494] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 622.024673] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.024821] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 622.024970] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 622.025269] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 622.025471] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 622.025597] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 622.025757] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 622.025928] env[61911]: DEBUG nova.virt.hardware [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 622.026991] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a86473-9fda-444c-b726-08bd2ebdb118 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.035532] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649eb151-b431-4836-9d0b-121e27695769 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.049163] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 622.054816] env[61911]: DEBUG oslo.service.loopingcall [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.055055] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 622.055413] env[61911]: INFO nova.compute.manager [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Rebuilding instance [ 622.057486] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62df95a9-5000-435d-8a11-08183d4b6a68 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.071105] env[61911]: DEBUG nova.scheduler.client.report [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 622.084734] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 622.084734] env[61911]: value = "task-1250876" [ 622.084734] env[61911]: _type = "Task" [ 622.084734] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.091981] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250876, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.108856] env[61911]: DEBUG nova.compute.manager [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 622.109740] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48738916-4508-46ba-9325-704919ab676f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.172022] env[61911]: INFO nova.compute.manager [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] [instance: 9f1af494-bb6c-4f7e-8218-0413848ceb5c] Took 1.03 seconds to deallocate network for instance. [ 622.514345] env[61911]: DEBUG nova.network.neutron [-] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.578255] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.578991] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 622.587230] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.555s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.604300] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250876, 'name': CreateVM_Task, 'duration_secs': 0.30458} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.604523] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 622.605088] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.605317] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.605751] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 622.606081] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e72e4529-5245-4b37-83c1-ada642c3d9bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.613027] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 622.613027] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526cb4d0-5b44-ad81-5183-136784caeb88" [ 622.613027] env[61911]: _type = "Task" [ 622.613027] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.625384] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526cb4d0-5b44-ad81-5183-136784caeb88, 'name': SearchDatastore_Task, 'duration_secs': 0.010017} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.626432] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.626771] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 622.626876] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.627222] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.627612] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 622.628384] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea525c74-7bc4-4498-8ff7-df2cee52ff6d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.637719] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 622.637922] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 622.639124] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae62de38-e64e-40b3-a84f-dbec465bd716 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.644641] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 622.644641] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c33b4c-5309-9fd3-0369-19eff231b5df" [ 622.644641] env[61911]: _type = "Task" [ 622.644641] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.652941] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c33b4c-5309-9fd3-0369-19eff231b5df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.018325] env[61911]: INFO nova.compute.manager [-] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Took 1.03 seconds to deallocate network for instance. [ 623.020236] env[61911]: DEBUG nova.compute.claims [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.020580] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.094177] env[61911]: DEBUG nova.compute.utils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 623.100190] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 623.100190] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 623.129926] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 623.131543] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-907684af-6e7c-4d05-bd7a-482c95aa7a1f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.138237] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 623.138237] env[61911]: value = "task-1250877" [ 623.138237] env[61911]: _type = "Task" [ 623.138237] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.151700] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.167256] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c33b4c-5309-9fd3-0369-19eff231b5df, 'name': SearchDatastore_Task, 'duration_secs': 0.007981} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.169117] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf84ab3e-d25c-4f3b-b3ac-a23b840604ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.175028] env[61911]: DEBUG nova.policy [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '781c6b5924124b198a9f0aa535419796', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8a3e350f52a473ea012eaccc7d44cd2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.177406] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 623.177406] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528a10bf-8a0e-feb9-a763-2483d9633112" [ 623.177406] env[61911]: _type = "Task" [ 623.177406] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.198124] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528a10bf-8a0e-feb9-a763-2483d9633112, 'name': SearchDatastore_Task, 'duration_secs': 0.009581} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.198124] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.198124] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 623.198124] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d3dd3c3-ac01-4fd0-9127-de542fc01e75 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.202928] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 623.202928] env[61911]: value = "task-1250878" [ 623.202928] env[61911]: _type = "Task" [ 623.202928] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.216407] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250878, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.228434] env[61911]: INFO nova.scheduler.client.report [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Deleted allocations for instance 9f1af494-bb6c-4f7e-8218-0413848ceb5c [ 623.602814] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 623.650329] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250877, 'name': PowerOffVM_Task, 'duration_secs': 0.123377} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.650626] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 623.650857] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 623.651660] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fade5740-9a23-487a-abc7-6e8db05bf83b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.662384] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 623.662384] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-172434de-8f2e-4e32-b4f6-015fcad0e8f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.670843] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd774672-dbda-4b68-9694-1a4ad031035f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.678818] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b4d138-6866-4ac0-8a62-057e99796879 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.716020] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811cb211-8610-454f-a663-681f0c528b1e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.717270] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 623.717399] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 623.717570] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Deleting the datastore file [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 623.718403] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-667bf421-6412-41e3-836a-18c4c2602e76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.727803] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104e7f24-cc2f-4d28-90fc-a37ace8c03ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.731771] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250878, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456204} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.734865] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Successfully created port: a377842f-2c4f-4b9f-9938-a33db11ae3b1 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.735922] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 623.736769] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 623.737202] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb050758-6107-4552-b2d0-8fc5d55eff7c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.750162] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb2a1517-0d84-4da6-b035-e2d45757ea03 tempest-VolumesAssistedSnapshotsTest-525324803 tempest-VolumesAssistedSnapshotsTest-525324803-project-member] Lock "9f1af494-bb6c-4f7e-8218-0413848ceb5c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.885s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.751119] env[61911]: DEBUG nova.compute.provider_tree [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.752357] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 623.752357] env[61911]: value = "task-1250880" [ 623.752357] env[61911]: _type = "Task" [ 623.752357] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.759181] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 623.759181] env[61911]: value = "task-1250881" [ 623.759181] env[61911]: _type = "Task" [ 623.759181] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.765328] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250880, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.771220] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250881, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.255912] env[61911]: DEBUG nova.scheduler.client.report [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 624.263117] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 624.277642] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250881, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066063} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.278167] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250880, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106684} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.278624] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 624.279046] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 624.279046] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 624.279046] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.283156] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8d7330-bf51-4ee5-94f8-94cee7ec1ebd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.309121] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 624.309753] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-408a14a2-712b-42e4-aa35-f9f2b94f5f93 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.333926] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 624.333926] env[61911]: value = "task-1250882" [ 624.333926] env[61911]: _type = "Task" [ 624.333926] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.343219] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250882, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.613862] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 624.651896] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 624.652149] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.652303] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 624.652481] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.652620] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 624.652761] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 624.653025] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 624.653416] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 624.653630] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 624.653809] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 624.653980] env[61911]: DEBUG nova.virt.hardware [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 624.659037] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b4b0b1-0b76-432b-b7d3-7e76d0a3474a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.664246] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8fd7b0-4f9d-49fb-917b-37c2bd31d5a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.745893] env[61911]: DEBUG nova.compute.manager [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Received event network-changed-a377842f-2c4f-4b9f-9938-a33db11ae3b1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 624.750620] env[61911]: DEBUG nova.compute.manager [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Refreshing instance network info cache due to event network-changed-a377842f-2c4f-4b9f-9938-a33db11ae3b1. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 624.750844] env[61911]: DEBUG oslo_concurrency.lockutils [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] Acquiring lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.750960] env[61911]: DEBUG oslo_concurrency.lockutils [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] Acquired lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.751208] env[61911]: DEBUG nova.network.neutron [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Refreshing network info cache for port a377842f-2c4f-4b9f-9938-a33db11ae3b1 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 624.764801] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Acquiring lock "b6773f2b-df86-4a4c-9b20-a314e81a7aa0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.765089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Lock "b6773f2b-df86-4a4c-9b20-a314e81a7aa0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.771036] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.184s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.771639] env[61911]: ERROR nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Traceback (most recent call last): [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self.driver.spawn(context, instance, image_meta, [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] vm_ref = self.build_virtual_machine(instance, [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.771639] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] for vif in network_info: [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return self._sync_wrapper(fn, *args, **kwargs) [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self.wait() [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self[:] = self._gt.wait() [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return self._exit_event.wait() [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] result = hub.switch() [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.771929] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return self.greenlet.switch() [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] result = function(*args, **kwargs) [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] return func(*args, **kwargs) [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] raise e [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] nwinfo = self.network_api.allocate_for_instance( [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] created_port_ids = self._update_ports_for_instance( [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] with excutils.save_and_reraise_exception(): [ 624.772210] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] self.force_reraise() [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] raise self.value [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] updated_port = self._update_port( [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] _ensure_no_port_binding_failure(port) [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] raise exception.PortBindingFailed(port_id=port['id']) [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] nova.exception.PortBindingFailed: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. [ 624.772486] env[61911]: ERROR nova.compute.manager [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] [ 624.773653] env[61911]: DEBUG nova.compute.utils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.780514] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.450s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.781757] env[61911]: INFO nova.compute.claims [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.785373] env[61911]: ERROR nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. [ 624.785373] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.785373] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.785373] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.785373] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.785373] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.785373] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.785373] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.785373] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.785373] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 624.785373] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.785373] env[61911]: ERROR nova.compute.manager raise self.value [ 624.785373] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.785373] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.785373] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.785373] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.785748] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.785748] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.785748] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. [ 624.785748] env[61911]: ERROR nova.compute.manager [ 624.785841] env[61911]: Traceback (most recent call last): [ 624.785868] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.785868] env[61911]: listener.cb(fileno) [ 624.785868] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.785868] env[61911]: result = function(*args, **kwargs) [ 624.785868] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.785868] env[61911]: return func(*args, **kwargs) [ 624.785868] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 624.785868] env[61911]: raise e [ 624.785868] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.785868] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 624.785868] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.785868] env[61911]: created_port_ids = self._update_ports_for_instance( [ 624.786141] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.786141] env[61911]: with excutils.save_and_reraise_exception(): [ 624.786141] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.786141] env[61911]: self.force_reraise() [ 624.786141] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.786141] env[61911]: raise self.value [ 624.786141] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.786141] env[61911]: updated_port = self._update_port( [ 624.786141] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.786141] env[61911]: _ensure_no_port_binding_failure(port) [ 624.786141] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.786141] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.786141] env[61911]: nova.exception.PortBindingFailed: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. [ 624.786141] env[61911]: Removing descriptor: 16 [ 624.786973] env[61911]: DEBUG nova.network.neutron [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.794030] env[61911]: ERROR nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Traceback (most recent call last): [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] yield resources [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self.driver.spawn(context, instance, image_meta, [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] vm_ref = self.build_virtual_machine(instance, [ 624.794030] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] for vif in network_info: [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return self._sync_wrapper(fn, *args, **kwargs) [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self.wait() [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self[:] = self._gt.wait() [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return self._exit_event.wait() [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.794739] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] result = hub.switch() [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return self.greenlet.switch() [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] result = function(*args, **kwargs) [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return func(*args, **kwargs) [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] raise e [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] nwinfo = self.network_api.allocate_for_instance( [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] created_port_ids = self._update_ports_for_instance( [ 624.795220] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] with excutils.save_and_reraise_exception(): [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self.force_reraise() [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] raise self.value [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] updated_port = self._update_port( [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] _ensure_no_port_binding_failure(port) [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] raise exception.PortBindingFailed(port_id=port['id']) [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] nova.exception.PortBindingFailed: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. [ 624.795538] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] [ 624.795802] env[61911]: INFO nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Terminating instance [ 624.805920] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Build of instance 2bb4e0d0-8154-45d8-b662-02e5a9c28b21 was re-scheduled: Binding failed for port 618f77ed-5e90-47b5-b69d-c7cb033c7f31, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 624.805920] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 624.805920] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Acquiring lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.805920] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Acquired lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.806136] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.816549] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.845515] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250882, 'name': ReconfigVM_Task, 'duration_secs': 0.267134} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.845831] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Reconfigured VM instance instance-00000011 to attach disk [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e/7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 624.846531] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-86ec0543-9de2-4c36-a283-83a28f68c93c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.852533] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 624.852533] env[61911]: value = "task-1250883" [ 624.852533] env[61911]: _type = "Task" [ 624.852533] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.866647] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250883, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.919876] env[61911]: DEBUG nova.network.neutron [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.302115] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Acquiring lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.323816] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 625.324103] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.324252] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 625.324436] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.324582] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 625.324728] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 625.324930] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 625.325102] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 625.325268] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 625.325428] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 625.325595] env[61911]: DEBUG nova.virt.hardware [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 625.326657] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec99eaca-b1f7-4149-99f3-b465391ac3bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.331141] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.338412] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b883a1-4348-4a9a-bd17-393228a927e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.351382] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 625.358187] env[61911]: DEBUG oslo.service.loopingcall [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.358617] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 625.361863] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ba06ba9-59ac-48b4-8e9f-65b0958ac86a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.379255] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250883, 'name': Rename_Task, 'duration_secs': 0.137555} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.380514] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 625.380834] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 625.380834] env[61911]: value = "task-1250884" [ 625.380834] env[61911]: _type = "Task" [ 625.380834] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.381086] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b8354b8-e485-4683-9048-d0fe5ed79c7b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.390251] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250884, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.392595] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 625.392595] env[61911]: value = "task-1250885" [ 625.392595] env[61911]: _type = "Task" [ 625.392595] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.400029] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250885, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.425700] env[61911]: DEBUG oslo_concurrency.lockutils [req-54a9f3cb-116a-42f0-977a-139c3341b53d req-2f610854-7f25-428f-bd09-908486a21838 service nova] Releasing lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.426110] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Acquired lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.426296] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.460318] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.894558] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250884, 'name': CreateVM_Task, 'duration_secs': 0.30436} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.898695] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 625.899141] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.899310] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.899615] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 625.900449] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85e5bc90-9a89-40c8-abd4-8bcdd9470294 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.904612] env[61911]: DEBUG oslo_vmware.api [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250885, 'name': PowerOnVM_Task, 'duration_secs': 0.43464} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.905228] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 625.905468] env[61911]: DEBUG nova.compute.manager [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 625.906249] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4908fe-906e-4b5c-bbb9-743b822efce3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.911999] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 625.911999] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d54b2d-c865-0f8b-1c85-1434b1092465" [ 625.911999] env[61911]: _type = "Task" [ 625.911999] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.928096] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d54b2d-c865-0f8b-1c85-1434b1092465, 'name': SearchDatastore_Task, 'duration_secs': 0.011122} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.930194] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.930841] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.930841] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.930841] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.931032] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.931318] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04dc5541-36ae-41a3-8e84-8d38e3923082 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.941108] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.941108] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 625.941108] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28e684fb-73ee-4437-88b1-4bd682decd70 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.944833] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 625.944833] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bb5608-b5f9-d5cf-99b1-80474c1cf065" [ 625.944833] env[61911]: _type = "Task" [ 625.944833] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.957391] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bb5608-b5f9-d5cf-99b1-80474c1cf065, 'name': SearchDatastore_Task, 'duration_secs': 0.009353} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.958627] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.961178] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-320947fc-523a-47c4-8dc9-01fed4c9a7d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.967084] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Releasing lock "refresh_cache-2bb4e0d0-8154-45d8-b662-02e5a9c28b21" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.967084] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 625.967084] env[61911]: DEBUG nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 625.967266] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.971657] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 625.971657] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52145d42-a4f4-23f5-82ac-6dc79e5fce2e" [ 625.971657] env[61911]: _type = "Task" [ 625.971657] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.979557] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52145d42-a4f4-23f5-82ac-6dc79e5fce2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.983321] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.213495] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d8d832-4042-4230-a762-547a3d345dcf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.221089] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2baf94d1-0cb2-4ad8-9c9a-706017f7a442 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.253933] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9a4d3a-2d7b-4c09-a508-7adcee5de4b0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.261830] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1ca47d-63b8-4acc-a39e-ba6bac4d05b6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.274762] env[61911]: DEBUG nova.compute.provider_tree [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.284387] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.434031] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.482265] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52145d42-a4f4-23f5-82ac-6dc79e5fce2e, 'name': SearchDatastore_Task, 'duration_secs': 0.016328} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.482540] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.482797] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 626.483494] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-979fc049-6b28-4882-adf6-8fb4767c617a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.485307] env[61911]: DEBUG nova.network.neutron [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.491310] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 626.491310] env[61911]: value = "task-1250886" [ 626.491310] env[61911]: _type = "Task" [ 626.491310] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.500169] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.778047] env[61911]: DEBUG nova.scheduler.client.report [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 626.786362] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Releasing lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.786783] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 626.786968] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.787274] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c0fcc5e-2fcd-4065-be44-09a4b50de9da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.796986] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a821f845-66c2-40bf-ba82-2bb4999ea08f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.811155] env[61911]: DEBUG nova.compute.manager [req-a9f8f384-a2a8-464b-9b6b-afda187e7bdd req-270302e3-e0ef-48c3-ba5d-e665b9998afc service nova] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Received event network-vif-deleted-a377842f-2c4f-4b9f-9938-a33db11ae3b1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 626.825650] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6d437f9-c341-4f4b-9b68-63711ca0ed42 could not be found. [ 626.825831] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.825929] env[61911]: INFO nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.826191] env[61911]: DEBUG oslo.service.loopingcall [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.826441] env[61911]: DEBUG nova.compute.manager [-] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 626.826540] env[61911]: DEBUG nova.network.neutron [-] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.864810] env[61911]: DEBUG nova.network.neutron [-] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.987669] env[61911]: INFO nova.compute.manager [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] [instance: 2bb4e0d0-8154-45d8-b662-02e5a9c28b21] Took 1.02 seconds to deallocate network for instance. [ 627.005027] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250886, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.083211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.083211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.083211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.083387] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.083566] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.085732] env[61911]: INFO nova.compute.manager [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Terminating instance [ 627.283397] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.284015] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 627.286722] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.787s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.339120] env[61911]: DEBUG nova.compute.manager [None req-fc053522-51ee-4357-abfb-9b7f5c61434b tempest-ServerDiagnosticsV248Test-1715709720 tempest-ServerDiagnosticsV248Test-1715709720-project-admin] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 627.339298] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b64b6a3-6b6d-4cad-8399-b6c9d12208f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.351565] env[61911]: INFO nova.compute.manager [None req-fc053522-51ee-4357-abfb-9b7f5c61434b tempest-ServerDiagnosticsV248Test-1715709720 tempest-ServerDiagnosticsV248Test-1715709720-project-admin] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Retrieving diagnostics [ 627.352445] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedaaa3c-fe7a-4202-9c2e-cbf46c7d8e1f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.385183] env[61911]: DEBUG nova.network.neutron [-] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.509564] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250886, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530827} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.509564] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 627.509944] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 627.509944] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6093302-f0f4-4f21-ba40-58d05275badc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.516028] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 627.516028] env[61911]: value = "task-1250887" [ 627.516028] env[61911]: _type = "Task" [ 627.516028] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.524394] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250887, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.590085] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "refresh_cache-7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.590275] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquired lock "refresh_cache-7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.590459] env[61911]: DEBUG nova.network.neutron [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.793656] env[61911]: DEBUG nova.compute.utils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.798939] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 627.802363] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.859867] env[61911]: DEBUG nova.policy [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08b6542498f34f038f46586b06823544', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dd9914e90fb4b52bf6222f4725f2101', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.891103] env[61911]: INFO nova.compute.manager [-] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Took 1.06 seconds to deallocate network for instance. [ 627.892277] env[61911]: DEBUG nova.compute.claims [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.892277] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.026302] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250887, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136471} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.029544] env[61911]: INFO nova.scheduler.client.report [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Deleted allocations for instance 2bb4e0d0-8154-45d8-b662-02e5a9c28b21 [ 628.036891] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 628.042160] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17434909-264f-4d7d-abe1-188177bf0d55 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.067428] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 628.071089] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b5f2375-145c-4e05-b968-d4ad28d34c15 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.095920] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 628.095920] env[61911]: value = "task-1250888" [ 628.095920] env[61911]: _type = "Task" [ 628.095920] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.740299] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 628.743894] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0fc8dd07-8db5-45c3-aba8-2f14543d6827 tempest-ServersWithSpecificFlavorTestJSON-681727852 tempest-ServersWithSpecificFlavorTestJSON-681727852-project-member] Lock "2bb4e0d0-8154-45d8-b662-02e5a9c28b21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.380s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.744167] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250888, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.744354] env[61911]: WARNING oslo_vmware.common.loopingcall [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] task run outlasted interval by 0.14794200000000002 sec [ 628.745220] env[61911]: DEBUG nova.network.neutron [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.752082] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Successfully created port: bdcb4666-4888-4dfc-9970-e4403d9f11cc {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.765894] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250888, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.841960] env[61911]: DEBUG nova.network.neutron [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.019252] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d72334-b3c3-4b35-826e-9d1630dff83f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.029614] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66ae3fc-0f96-4445-8d63-dbbe383056af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.068729] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30498e39-c15f-4df9-931b-540839aa9367 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.076263] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadd80da-6d3e-4ef6-8e5c-63f775fc8c31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.094284] env[61911]: DEBUG nova.compute.provider_tree [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.255612] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.255808] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.256065] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.256271] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.256436] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.262009] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 629.268028] env[61911]: INFO nova.compute.manager [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Terminating instance [ 629.272778] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250888, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.345268] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Releasing lock "refresh_cache-7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.348297] env[61911]: DEBUG nova.compute.manager [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 629.348297] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.348297] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eaee1cc-16d5-48bc-a12b-cda23461ddcf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.355949] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 629.356377] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c416895-5830-4035-9406-10923764a7f2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.363112] env[61911]: DEBUG oslo_vmware.api [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 629.363112] env[61911]: value = "task-1250889" [ 629.363112] env[61911]: _type = "Task" [ 629.363112] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.372384] env[61911]: DEBUG oslo_vmware.api [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.598181] env[61911]: DEBUG nova.scheduler.client.report [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.760760] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 629.771079] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "refresh_cache-90e2a5db-918a-4f66-a9c8-ef41dc4b855b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.771079] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquired lock "refresh_cache-90e2a5db-918a-4f66-a9c8-ef41dc4b855b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.771079] env[61911]: DEBUG nova.network.neutron [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.771849] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250888, 'name': ReconfigVM_Task, 'duration_secs': 1.41771} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.774081] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383/821d3f6f-e847-4ddb-ac00-ea55af302383.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.775013] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fff4aaf-cf0b-4d66-9189-98006a25d571 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.786298] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 629.786298] env[61911]: value = "task-1250890" [ 629.786298] env[61911]: _type = "Task" [ 629.786298] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.799356] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250890, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.800306] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.803972] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 629.804348] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.804348] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 629.804718] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.804718] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 629.804962] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 629.805451] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 629.806212] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 629.806212] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 629.806212] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 629.806610] env[61911]: DEBUG nova.virt.hardware [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 629.807312] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85e7a85-286c-4abc-b4ee-4d57793e1972 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.816203] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80acce53-35a2-463a-ae55-bd62ad93e3ca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.880833] env[61911]: DEBUG oslo_vmware.api [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250889, 'name': PowerOffVM_Task, 'duration_secs': 0.103854} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.883275] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 629.883442] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 629.883763] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9f81e04-8bf5-4a3d-b494-f13020ad2cdd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.908855] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 629.909389] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 629.909675] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Deleting the datastore file [datastore1] 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 629.910510] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd77d56f-28a4-45f6-8f18-3ca15bb3084e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.917473] env[61911]: DEBUG oslo_vmware.api [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for the task: (returnval){ [ 629.917473] env[61911]: value = "task-1250892" [ 629.917473] env[61911]: _type = "Task" [ 629.917473] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.927590] env[61911]: DEBUG oslo_vmware.api [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250892, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.040150] env[61911]: DEBUG nova.compute.manager [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Received event network-changed-bdcb4666-4888-4dfc-9970-e4403d9f11cc {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 630.040309] env[61911]: DEBUG nova.compute.manager [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Refreshing instance network info cache due to event network-changed-bdcb4666-4888-4dfc-9970-e4403d9f11cc. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 630.040528] env[61911]: DEBUG oslo_concurrency.lockutils [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] Acquiring lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.040670] env[61911]: DEBUG oslo_concurrency.lockutils [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] Acquired lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.040830] env[61911]: DEBUG nova.network.neutron [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Refreshing network info cache for port bdcb4666-4888-4dfc-9970-e4403d9f11cc {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 630.109125] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.820s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.109125] env[61911]: ERROR nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. [ 630.109125] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Traceback (most recent call last): [ 630.109125] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 630.109125] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self.driver.spawn(context, instance, image_meta, [ 630.109125] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 630.109125] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.109125] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.109125] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] vm_ref = self.build_virtual_machine(instance, [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] for vif in network_info: [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return self._sync_wrapper(fn, *args, **kwargs) [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self.wait() [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self[:] = self._gt.wait() [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return self._exit_event.wait() [ 630.109388] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] result = hub.switch() [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return self.greenlet.switch() [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] result = function(*args, **kwargs) [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] return func(*args, **kwargs) [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] raise e [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] nwinfo = self.network_api.allocate_for_instance( [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.109761] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] created_port_ids = self._update_ports_for_instance( [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] with excutils.save_and_reraise_exception(): [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] self.force_reraise() [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] raise self.value [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] updated_port = self._update_port( [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] _ensure_no_port_binding_failure(port) [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.110197] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] raise exception.PortBindingFailed(port_id=port['id']) [ 630.110526] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] nova.exception.PortBindingFailed: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. [ 630.110526] env[61911]: ERROR nova.compute.manager [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] [ 630.111503] env[61911]: DEBUG nova.compute.utils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.112795] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.569s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.114325] env[61911]: INFO nova.compute.claims [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.117548] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Build of instance 8da2407a-c7e5-437a-ab71-57b92d98182f was re-scheduled: Binding failed for port 0b3c0ab0-0be3-4503-b5ad-159f20c14904, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 630.118145] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 630.118461] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Acquiring lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.118715] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Acquired lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.118972] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.191356] env[61911]: ERROR nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. [ 630.191356] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.191356] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 630.191356] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.191356] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.191356] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.191356] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.191356] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.191356] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.191356] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 630.191356] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.191356] env[61911]: ERROR nova.compute.manager raise self.value [ 630.191356] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.191356] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.191356] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.191356] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.191926] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.191926] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.191926] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. [ 630.191926] env[61911]: ERROR nova.compute.manager [ 630.191926] env[61911]: Traceback (most recent call last): [ 630.191926] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.191926] env[61911]: listener.cb(fileno) [ 630.191926] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.191926] env[61911]: result = function(*args, **kwargs) [ 630.191926] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.191926] env[61911]: return func(*args, **kwargs) [ 630.191926] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 630.191926] env[61911]: raise e [ 630.191926] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 630.191926] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 630.191926] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.191926] env[61911]: created_port_ids = self._update_ports_for_instance( [ 630.191926] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.191926] env[61911]: with excutils.save_and_reraise_exception(): [ 630.191926] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.191926] env[61911]: self.force_reraise() [ 630.191926] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.191926] env[61911]: raise self.value [ 630.191926] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.191926] env[61911]: updated_port = self._update_port( [ 630.191926] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.191926] env[61911]: _ensure_no_port_binding_failure(port) [ 630.191926] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.191926] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.192678] env[61911]: nova.exception.PortBindingFailed: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. [ 630.192678] env[61911]: Removing descriptor: 19 [ 630.192678] env[61911]: ERROR nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Traceback (most recent call last): [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] yield resources [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self.driver.spawn(context, instance, image_meta, [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.192678] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] vm_ref = self.build_virtual_machine(instance, [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] for vif in network_info: [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return self._sync_wrapper(fn, *args, **kwargs) [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self.wait() [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self[:] = self._gt.wait() [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return self._exit_event.wait() [ 630.193012] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] result = hub.switch() [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return self.greenlet.switch() [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] result = function(*args, **kwargs) [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return func(*args, **kwargs) [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] raise e [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] nwinfo = self.network_api.allocate_for_instance( [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.193354] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] created_port_ids = self._update_ports_for_instance( [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] with excutils.save_and_reraise_exception(): [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self.force_reraise() [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] raise self.value [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] updated_port = self._update_port( [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] _ensure_no_port_binding_failure(port) [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.193677] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] raise exception.PortBindingFailed(port_id=port['id']) [ 630.193962] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] nova.exception.PortBindingFailed: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. [ 630.193962] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] [ 630.193962] env[61911]: INFO nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Terminating instance [ 630.293265] env[61911]: DEBUG nova.network.neutron [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.298236] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250890, 'name': Rename_Task, 'duration_secs': 0.144831} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.298521] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 630.298694] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65c11d13-5139-4e02-9794-13ead677a8c3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.304612] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Waiting for the task: (returnval){ [ 630.304612] env[61911]: value = "task-1250893" [ 630.304612] env[61911]: _type = "Task" [ 630.304612] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.314545] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.377939] env[61911]: DEBUG nova.network.neutron [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.428166] env[61911]: DEBUG oslo_vmware.api [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Task: {'id': task-1250892, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089854} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.428446] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 630.428628] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 630.428800] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 630.428970] env[61911]: INFO nova.compute.manager [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Took 1.08 seconds to destroy the instance on the hypervisor. [ 630.429228] env[61911]: DEBUG oslo.service.loopingcall [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.429411] env[61911]: DEBUG nova.compute.manager [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 630.429504] env[61911]: DEBUG nova.network.neutron [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.453991] env[61911]: DEBUG nova.network.neutron [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.567202] env[61911]: DEBUG nova.network.neutron [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.645162] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.672843] env[61911]: DEBUG nova.network.neutron [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.696155] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Acquiring lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.816319] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250893, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.823025] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.880173] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Releasing lock "refresh_cache-90e2a5db-918a-4f66-a9c8-ef41dc4b855b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.880589] env[61911]: DEBUG nova.compute.manager [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 630.880805] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 630.881662] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abf2c0d-5109-425c-a527-4fd985f936f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.890462] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 630.891446] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-475b3808-48c7-4cf1-b741-a41f32b084a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.898108] env[61911]: DEBUG oslo_vmware.api [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 630.898108] env[61911]: value = "task-1250894" [ 630.898108] env[61911]: _type = "Task" [ 630.898108] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.907389] env[61911]: DEBUG oslo_vmware.api [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250894, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.956213] env[61911]: DEBUG nova.network.neutron [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.180295] env[61911]: DEBUG oslo_concurrency.lockutils [req-5ac1bcbb-0fb4-48d8-9d98-33f9ea25c746 req-189b05d4-5851-418b-9fd7-669d6fc7672b service nova] Releasing lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.180777] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Acquired lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.181510] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.319105] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250893, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.325887] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Releasing lock "refresh_cache-8da2407a-c7e5-437a-ab71-57b92d98182f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.326013] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 631.326198] env[61911]: DEBUG nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 631.326363] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.353889] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.410051] env[61911]: DEBUG oslo_vmware.api [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250894, 'name': PowerOffVM_Task, 'duration_secs': 0.106181} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.410345] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 631.410538] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 631.410968] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0aadf0f8-4f08-4833-8191-6f40b3c9f58d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.431858] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 631.432089] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 631.432337] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Deleting the datastore file [datastore2] 90e2a5db-918a-4f66-a9c8-ef41dc4b855b {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 631.432575] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30dace69-e2f9-4ebf-b527-8c9c97921200 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.446011] env[61911]: DEBUG oslo_vmware.api [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for the task: (returnval){ [ 631.446011] env[61911]: value = "task-1250896" [ 631.446011] env[61911]: _type = "Task" [ 631.446011] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.458415] env[61911]: DEBUG oslo_vmware.api [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.461940] env[61911]: INFO nova.compute.manager [-] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Took 1.03 seconds to deallocate network for instance. [ 631.620255] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d0404b-ff40-4ff5-8213-8e5055bebec3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.630311] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b3e1fe-cd2c-4fb6-80c8-dd71dd827d8c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.670857] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0179f338-297a-465b-8dd8-46d3b9b1e1e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.679172] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f40872-6dcd-41ae-9f2e-de85e14cb3e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.697878] env[61911]: DEBUG nova.compute.provider_tree [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.708624] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.817698] env[61911]: DEBUG oslo_vmware.api [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Task: {'id': task-1250893, 'name': PowerOnVM_Task, 'duration_secs': 1.015325} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.817985] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 631.818208] env[61911]: DEBUG nova.compute.manager [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 631.819067] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aedd825-1afa-4ad8-aa6a-79a8115899d8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.841373] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.859139] env[61911]: DEBUG nova.network.neutron [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.956011] env[61911]: DEBUG oslo_vmware.api [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Task: {'id': task-1250896, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086854} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.957962] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 631.958404] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 631.958509] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.958751] env[61911]: INFO nova.compute.manager [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Took 1.08 seconds to destroy the instance on the hypervisor. [ 631.959058] env[61911]: DEBUG oslo.service.loopingcall [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.959272] env[61911]: DEBUG nova.compute.manager [-] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 631.959367] env[61911]: DEBUG nova.network.neutron [-] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.970227] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.979816] env[61911]: DEBUG nova.network.neutron [-] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.068291] env[61911]: DEBUG nova.compute.manager [req-5fc46e6d-11b5-487e-b1c1-573faac42dd2 req-c7f55b38-67d0-4d93-9eec-07374cd8fff1 service nova] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Received event network-vif-deleted-bdcb4666-4888-4dfc-9970-e4403d9f11cc {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 632.202675] env[61911]: DEBUG nova.scheduler.client.report [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 632.337316] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.344047] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Releasing lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.344412] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 632.344602] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.344875] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-554c1606-c935-4676-90a3-4ef5f193b959 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.354031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f020852c-082a-4036-bb4a-05f15ffb778b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.365468] env[61911]: INFO nova.compute.manager [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] [instance: 8da2407a-c7e5-437a-ab71-57b92d98182f] Took 1.04 seconds to deallocate network for instance. [ 632.379659] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95733825-2618-4f6a-b2eb-bdef3a7f60de could not be found. [ 632.379930] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.380134] env[61911]: INFO nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Took 0.04 seconds to destroy the instance on the hypervisor. [ 632.380400] env[61911]: DEBUG oslo.service.loopingcall [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.380636] env[61911]: DEBUG nova.compute.manager [-] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 632.380730] env[61911]: DEBUG nova.network.neutron [-] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.399676] env[61911]: DEBUG nova.network.neutron [-] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.481891] env[61911]: DEBUG nova.network.neutron [-] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.708900] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.709440] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 632.713397] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.618s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.714865] env[61911]: INFO nova.compute.claims [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.902261] env[61911]: DEBUG nova.network.neutron [-] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.988250] env[61911]: INFO nova.compute.manager [-] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Took 1.03 seconds to deallocate network for instance. [ 633.221311] env[61911]: DEBUG nova.compute.utils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.222679] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 633.222852] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 633.287945] env[61911]: DEBUG nova.policy [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '78f28c7011734b7d8b906de60dc00679', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '797c1fd4dec749f3ba3b144ab57af983', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.405235] env[61911]: INFO nova.scheduler.client.report [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Deleted allocations for instance 8da2407a-c7e5-437a-ab71-57b92d98182f [ 633.411232] env[61911]: INFO nova.compute.manager [-] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Took 1.03 seconds to deallocate network for instance. [ 633.419436] env[61911]: DEBUG nova.compute.claims [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.419632] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.497256] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.727911] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 633.799242] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "821d3f6f-e847-4ddb-ac00-ea55af302383" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.799502] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "821d3f6f-e847-4ddb-ac00-ea55af302383" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.799679] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "821d3f6f-e847-4ddb-ac00-ea55af302383-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.799916] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "821d3f6f-e847-4ddb-ac00-ea55af302383-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.800905] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "821d3f6f-e847-4ddb-ac00-ea55af302383-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.803620] env[61911]: INFO nova.compute.manager [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Terminating instance [ 633.900867] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Successfully created port: edca9748-9f39-4d6c-906e-3771d1540787 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.918189] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61874348-c732-4373-b43e-12d6fe9bf86f tempest-FloatingIPsAssociationTestJSON-1101347919 tempest-FloatingIPsAssociationTestJSON-1101347919-project-member] Lock "8da2407a-c7e5-437a-ab71-57b92d98182f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.711s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.185657] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83728cd6-5b26-40a3-bec5-dfcd6c0bd146 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.193307] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5ab508-c66d-461d-b0c5-ba9f5cbedb0f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.228583] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764bc651-993b-419b-8e08-dde2a5ff3c0f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.240846] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa1c8bc-144b-4aa0-b2e4-44c22761d4dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.257980] env[61911]: DEBUG nova.compute.provider_tree [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.311611] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "refresh_cache-821d3f6f-e847-4ddb-ac00-ea55af302383" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.311790] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquired lock "refresh_cache-821d3f6f-e847-4ddb-ac00-ea55af302383" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.311964] env[61911]: DEBUG nova.network.neutron [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.421535] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 634.748858] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 634.760420] env[61911]: DEBUG nova.scheduler.client.report [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 634.779025] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 634.780440] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.780618] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 634.780807] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.781534] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 634.784215] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 634.785160] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 634.785160] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 634.785160] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 634.785160] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 634.785160] env[61911]: DEBUG nova.virt.hardware [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 634.786325] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b37163-386b-4a5f-ad36-a48d4306ea08 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.795807] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a72b2b-26fa-4da2-a796-d33b1ab6d3a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.844571] env[61911]: DEBUG nova.network.neutron [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.946032] env[61911]: DEBUG nova.network.neutron [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.959577] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.265569] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.265786] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 635.271528] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.066s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.271528] env[61911]: INFO nova.compute.claims [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.292205] env[61911]: DEBUG nova.compute.manager [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Received event network-changed-edca9748-9f39-4d6c-906e-3771d1540787 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 635.292429] env[61911]: DEBUG nova.compute.manager [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Refreshing instance network info cache due to event network-changed-edca9748-9f39-4d6c-906e-3771d1540787. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 635.292646] env[61911]: DEBUG oslo_concurrency.lockutils [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] Acquiring lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.296020] env[61911]: DEBUG oslo_concurrency.lockutils [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] Acquired lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.296020] env[61911]: DEBUG nova.network.neutron [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Refreshing network info cache for port edca9748-9f39-4d6c-906e-3771d1540787 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 635.451211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Releasing lock "refresh_cache-821d3f6f-e847-4ddb-ac00-ea55af302383" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.451914] env[61911]: DEBUG nova.compute.manager [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 635.452167] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 635.453029] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254caadf-b101-4e99-8417-f31c679b9ab2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.462590] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 635.462956] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67ea11b6-7ee3-431b-8522-c31c400d8595 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.474020] env[61911]: DEBUG oslo_vmware.api [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 635.474020] env[61911]: value = "task-1250897" [ 635.474020] env[61911]: _type = "Task" [ 635.474020] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.482840] env[61911]: DEBUG oslo_vmware.api [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250897, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.607458] env[61911]: ERROR nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. [ 635.607458] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.607458] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.607458] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.607458] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.607458] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.607458] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.607458] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.607458] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.607458] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 635.607458] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.607458] env[61911]: ERROR nova.compute.manager raise self.value [ 635.607458] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.607458] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.607458] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.607458] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.607889] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.607889] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.607889] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. [ 635.607889] env[61911]: ERROR nova.compute.manager [ 635.607889] env[61911]: Traceback (most recent call last): [ 635.607889] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.607889] env[61911]: listener.cb(fileno) [ 635.607889] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.607889] env[61911]: result = function(*args, **kwargs) [ 635.607889] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.607889] env[61911]: return func(*args, **kwargs) [ 635.607889] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 635.607889] env[61911]: raise e [ 635.607889] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.607889] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 635.607889] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.607889] env[61911]: created_port_ids = self._update_ports_for_instance( [ 635.607889] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.607889] env[61911]: with excutils.save_and_reraise_exception(): [ 635.607889] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.607889] env[61911]: self.force_reraise() [ 635.607889] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.607889] env[61911]: raise self.value [ 635.607889] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.607889] env[61911]: updated_port = self._update_port( [ 635.607889] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.607889] env[61911]: _ensure_no_port_binding_failure(port) [ 635.607889] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.607889] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.608538] env[61911]: nova.exception.PortBindingFailed: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. [ 635.608538] env[61911]: Removing descriptor: 16 [ 635.608538] env[61911]: ERROR nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Traceback (most recent call last): [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] yield resources [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self.driver.spawn(context, instance, image_meta, [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.608538] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] vm_ref = self.build_virtual_machine(instance, [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] for vif in network_info: [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return self._sync_wrapper(fn, *args, **kwargs) [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self.wait() [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self[:] = self._gt.wait() [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return self._exit_event.wait() [ 635.608819] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] result = hub.switch() [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return self.greenlet.switch() [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] result = function(*args, **kwargs) [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return func(*args, **kwargs) [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] raise e [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] nwinfo = self.network_api.allocate_for_instance( [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.609096] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] created_port_ids = self._update_ports_for_instance( [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] with excutils.save_and_reraise_exception(): [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self.force_reraise() [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] raise self.value [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] updated_port = self._update_port( [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] _ensure_no_port_binding_failure(port) [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.609375] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] raise exception.PortBindingFailed(port_id=port['id']) [ 635.609690] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] nova.exception.PortBindingFailed: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. [ 635.609690] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] [ 635.609690] env[61911]: INFO nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Terminating instance [ 635.777594] env[61911]: DEBUG nova.compute.utils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.781549] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 635.781695] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 635.817313] env[61911]: DEBUG nova.network.neutron [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.850466] env[61911]: DEBUG nova.policy [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'baa9ad2d71b9447abcb7460d9a35af8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d63d1c52b244418895f3d1cc3f77175', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.931985] env[61911]: DEBUG nova.network.neutron [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.984057] env[61911]: DEBUG oslo_vmware.api [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250897, 'name': PowerOffVM_Task, 'duration_secs': 0.197448} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.984354] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 635.984524] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 635.984768] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b80d759-718c-41c6-828b-dc602f32d4dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.011173] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 636.012824] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 636.012824] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleting the datastore file [datastore1] 821d3f6f-e847-4ddb-ac00-ea55af302383 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 636.012824] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-187f43f9-fe95-4b3c-ae74-64991525c770 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.020132] env[61911]: DEBUG oslo_vmware.api [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for the task: (returnval){ [ 636.020132] env[61911]: value = "task-1250899" [ 636.020132] env[61911]: _type = "Task" [ 636.020132] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.032017] env[61911]: DEBUG oslo_vmware.api [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.116393] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Acquiring lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.227458] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Successfully created port: 208959bd-3a8b-489d-8a55-9ed7aedf7055 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.288053] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 636.435858] env[61911]: DEBUG oslo_concurrency.lockutils [req-a9c1f767-9d46-45e6-8794-d3cfcb8aede2 req-d2182469-8437-41c9-a0c4-1c39e7d36bd9 service nova] Releasing lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.439020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Acquired lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.439020] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.534825] env[61911]: DEBUG oslo_vmware.api [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Task: {'id': task-1250899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101095} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.534825] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 636.534825] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 636.534825] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.535094] env[61911]: INFO nova.compute.manager [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Took 1.08 seconds to destroy the instance on the hypervisor. [ 636.535465] env[61911]: DEBUG oslo.service.loopingcall [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.535581] env[61911]: DEBUG nova.compute.manager [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 636.535918] env[61911]: DEBUG nova.network.neutron [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 636.556199] env[61911]: DEBUG nova.network.neutron [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.793220] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be90ad25-1545-4e26-acab-98e7f7c0fe76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.805919] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f355c6a9-5545-4c9e-affa-959a7eeee1c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.847330] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81c6354-46c8-4e9b-b33e-4bc0f76e66ca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.856580] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7d66c2-968b-4232-aee3-9204007a116f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.871985] env[61911]: DEBUG nova.compute.provider_tree [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.962293] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.061864] env[61911]: DEBUG nova.network.neutron [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.072925] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.307699] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 637.327752] env[61911]: DEBUG nova.compute.manager [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Received event network-vif-deleted-edca9748-9f39-4d6c-906e-3771d1540787 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 637.328752] env[61911]: DEBUG nova.compute.manager [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Received event network-changed-208959bd-3a8b-489d-8a55-9ed7aedf7055 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 637.328752] env[61911]: DEBUG nova.compute.manager [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Refreshing instance network info cache due to event network-changed-208959bd-3a8b-489d-8a55-9ed7aedf7055. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 637.329588] env[61911]: DEBUG oslo_concurrency.lockutils [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] Acquiring lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.329588] env[61911]: DEBUG oslo_concurrency.lockutils [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] Acquired lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.329588] env[61911]: DEBUG nova.network.neutron [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Refreshing network info cache for port 208959bd-3a8b-489d-8a55-9ed7aedf7055 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.344807] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 637.345049] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.345217] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 637.345403] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.345549] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 637.345699] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 637.345906] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 637.346080] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 637.346249] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 637.346412] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 637.346585] env[61911]: DEBUG nova.virt.hardware [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 637.347789] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfbf91c-a5df-48aa-90dc-23248a5b13b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.356160] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2141b5-85ca-4817-aac8-26d55914594c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.376551] env[61911]: DEBUG nova.scheduler.client.report [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 637.565428] env[61911]: INFO nova.compute.manager [-] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Took 1.03 seconds to deallocate network for instance. [ 637.575600] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Releasing lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.577116] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 637.577116] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.577397] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-773b7974-b6ea-492d-8a48-ccb29ebba5f2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.588908] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29903195-7f43-4753-8749-4ec4b6207229 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.603605] env[61911]: ERROR nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. [ 637.603605] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.603605] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.603605] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.603605] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.603605] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.603605] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.603605] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.603605] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.603605] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 637.603605] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.603605] env[61911]: ERROR nova.compute.manager raise self.value [ 637.603605] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.603605] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.603605] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.603605] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.604024] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.604024] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.604024] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. [ 637.604024] env[61911]: ERROR nova.compute.manager [ 637.604024] env[61911]: Traceback (most recent call last): [ 637.604024] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.604024] env[61911]: listener.cb(fileno) [ 637.604024] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.604024] env[61911]: result = function(*args, **kwargs) [ 637.604024] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.604024] env[61911]: return func(*args, **kwargs) [ 637.604024] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.604024] env[61911]: raise e [ 637.604024] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.604024] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 637.604024] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.604024] env[61911]: created_port_ids = self._update_ports_for_instance( [ 637.604024] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.604024] env[61911]: with excutils.save_and_reraise_exception(): [ 637.604024] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.604024] env[61911]: self.force_reraise() [ 637.604024] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.604024] env[61911]: raise self.value [ 637.604024] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.604024] env[61911]: updated_port = self._update_port( [ 637.604024] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.604024] env[61911]: _ensure_no_port_binding_failure(port) [ 637.604024] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.604024] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.605020] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. [ 637.605020] env[61911]: Removing descriptor: 16 [ 637.605020] env[61911]: ERROR nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Traceback (most recent call last): [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] yield resources [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self.driver.spawn(context, instance, image_meta, [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.605020] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] vm_ref = self.build_virtual_machine(instance, [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] for vif in network_info: [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return self._sync_wrapper(fn, *args, **kwargs) [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self.wait() [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self[:] = self._gt.wait() [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return self._exit_event.wait() [ 637.605297] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] result = hub.switch() [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return self.greenlet.switch() [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] result = function(*args, **kwargs) [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return func(*args, **kwargs) [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] raise e [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] nwinfo = self.network_api.allocate_for_instance( [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.605582] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] created_port_ids = self._update_ports_for_instance( [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] with excutils.save_and_reraise_exception(): [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self.force_reraise() [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] raise self.value [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] updated_port = self._update_port( [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] _ensure_no_port_binding_failure(port) [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.605861] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] raise exception.PortBindingFailed(port_id=port['id']) [ 637.606136] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] nova.exception.PortBindingFailed: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. [ 637.606136] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] [ 637.606136] env[61911]: INFO nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Terminating instance [ 637.618262] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 17eec473-7082-4103-baf5-b61dc0d63281 could not be found. [ 637.618495] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.618673] env[61911]: INFO nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Took 0.04 seconds to destroy the instance on the hypervisor. [ 637.618914] env[61911]: DEBUG oslo.service.loopingcall [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.620050] env[61911]: DEBUG nova.compute.manager [-] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 637.620156] env[61911]: DEBUG nova.network.neutron [-] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.637057] env[61911]: DEBUG nova.network.neutron [-] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.853030] env[61911]: DEBUG nova.network.neutron [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.882469] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.882995] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 637.886117] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.305s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.939800] env[61911]: DEBUG nova.network.neutron [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.077205] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.110506] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Acquiring lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.139181] env[61911]: DEBUG nova.network.neutron [-] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.396505] env[61911]: DEBUG nova.compute.utils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.398021] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 638.398189] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 638.442676] env[61911]: DEBUG oslo_concurrency.lockutils [req-27138da1-ab0d-42e2-93dc-2ff177ce638b req-26b45335-bff9-4d88-bc5c-72feab9163c8 service nova] Releasing lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.445208] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Acquired lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.445208] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.449285] env[61911]: DEBUG nova.policy [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '756c990ab59d41e491e9a181ee28597d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '14c17884ee544b4493db3243cac72c16', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 638.645444] env[61911]: INFO nova.compute.manager [-] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Took 1.03 seconds to deallocate network for instance. [ 638.651161] env[61911]: DEBUG nova.compute.claims [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.651334] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.852487] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50c0938-7ba6-457a-a108-9cd65b2ac2d1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.861111] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68517866-bfc3-4176-9a51-5fe9349333bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.897010] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2da6c6-0b1c-4ea3-aba1-10f52edbcbd5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.904882] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 638.910632] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8452b3c5-b7d1-41df-a514-0851cdd7ad5e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.926901] env[61911]: DEBUG nova.compute.provider_tree [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.968943] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.024144] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Successfully created port: 38f28b29-a1fa-4315-9642-5604bd7c3346 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.143346] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.433102] env[61911]: DEBUG nova.scheduler.client.report [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 639.577783] env[61911]: DEBUG nova.compute.manager [req-c6477d46-f892-4de7-93bd-ae8bc582971f req-6b276005-330e-4c4b-93cf-749f6b768152 service nova] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Received event network-vif-deleted-208959bd-3a8b-489d-8a55-9ed7aedf7055 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 639.645957] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Releasing lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.646470] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 639.646828] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.646977] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9020001-249c-4fd5-b6a1-99c6d2d12e40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.657242] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317b1e14-e9a7-4a66-a704-fda1e1a1a8d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.688163] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91e485bf-4751-4820-8f77-47bd79cc8423 could not be found. [ 639.688411] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.688612] env[61911]: INFO nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Took 0.04 seconds to destroy the instance on the hypervisor. [ 639.688867] env[61911]: DEBUG oslo.service.loopingcall [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.689179] env[61911]: DEBUG nova.compute.manager [-] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 639.689275] env[61911]: DEBUG nova.network.neutron [-] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.721556] env[61911]: DEBUG nova.network.neutron [-] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.921194] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 639.937731] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.937731] env[61911]: ERROR nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. [ 639.937731] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Traceback (most recent call last): [ 639.937731] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 639.937731] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self.driver.spawn(context, instance, image_meta, [ 639.937731] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 639.937731] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.937731] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.937731] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] vm_ref = self.build_virtual_machine(instance, [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] for vif in network_info: [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return self._sync_wrapper(fn, *args, **kwargs) [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self.wait() [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self[:] = self._gt.wait() [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return self._exit_event.wait() [ 639.938086] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] result = hub.switch() [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return self.greenlet.switch() [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] result = function(*args, **kwargs) [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] return func(*args, **kwargs) [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] raise e [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] nwinfo = self.network_api.allocate_for_instance( [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.938417] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] created_port_ids = self._update_ports_for_instance( [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] with excutils.save_and_reraise_exception(): [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] self.force_reraise() [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] raise self.value [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] updated_port = self._update_port( [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] _ensure_no_port_binding_failure(port) [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.938732] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] raise exception.PortBindingFailed(port_id=port['id']) [ 639.939040] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] nova.exception.PortBindingFailed: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. [ 639.939040] env[61911]: ERROR nova.compute.manager [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] [ 639.939040] env[61911]: DEBUG nova.compute.utils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 639.941796] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.335s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.942070] env[61911]: DEBUG nova.objects.instance [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 639.944995] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Build of instance c4a4e40a-c3d3-4d27-98a4-452f18c6408d was re-scheduled: Binding failed for port 2371d3f8-8347-4e79-abd6-4d5882113fb0, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 639.949658] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 639.949926] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.950098] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.950266] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 639.960871] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 639.961190] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.961190] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 639.961254] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.961414] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 639.961545] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 639.961743] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 639.962195] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 639.962368] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 639.962538] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 639.962705] env[61911]: DEBUG nova.virt.hardware [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 639.963591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547f92a3-67c8-4ea8-8ed1-c9e185aa7a45 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.977939] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75b942d-625d-4fad-b8cf-c96c9b94e4b4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.125533] env[61911]: ERROR nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. [ 640.125533] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.125533] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.125533] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.125533] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.125533] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.125533] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.125533] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.125533] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.125533] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 640.125533] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.125533] env[61911]: ERROR nova.compute.manager raise self.value [ 640.125533] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.125533] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.125533] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.125533] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.125953] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.125953] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.125953] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. [ 640.125953] env[61911]: ERROR nova.compute.manager [ 640.126310] env[61911]: Traceback (most recent call last): [ 640.128019] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.128019] env[61911]: listener.cb(fileno) [ 640.128019] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.128019] env[61911]: result = function(*args, **kwargs) [ 640.128019] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.128019] env[61911]: return func(*args, **kwargs) [ 640.128019] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.128019] env[61911]: raise e [ 640.128019] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.128019] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 640.128019] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.128019] env[61911]: created_port_ids = self._update_ports_for_instance( [ 640.128019] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.128019] env[61911]: with excutils.save_and_reraise_exception(): [ 640.128019] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.128019] env[61911]: self.force_reraise() [ 640.128019] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.128019] env[61911]: raise self.value [ 640.128019] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.128019] env[61911]: updated_port = self._update_port( [ 640.128019] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.128019] env[61911]: _ensure_no_port_binding_failure(port) [ 640.128019] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.128019] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.128019] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. [ 640.128019] env[61911]: Removing descriptor: 16 [ 640.128790] env[61911]: ERROR nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Traceback (most recent call last): [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] yield resources [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self.driver.spawn(context, instance, image_meta, [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] vm_ref = self.build_virtual_machine(instance, [ 640.128790] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] for vif in network_info: [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return self._sync_wrapper(fn, *args, **kwargs) [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self.wait() [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self[:] = self._gt.wait() [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return self._exit_event.wait() [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.129193] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] result = hub.switch() [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return self.greenlet.switch() [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] result = function(*args, **kwargs) [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return func(*args, **kwargs) [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] raise e [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] nwinfo = self.network_api.allocate_for_instance( [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] created_port_ids = self._update_ports_for_instance( [ 640.129484] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] with excutils.save_and_reraise_exception(): [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self.force_reraise() [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] raise self.value [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] updated_port = self._update_port( [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] _ensure_no_port_binding_failure(port) [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] raise exception.PortBindingFailed(port_id=port['id']) [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] nova.exception.PortBindingFailed: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. [ 640.129907] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] [ 640.130252] env[61911]: INFO nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Terminating instance [ 640.227136] env[61911]: DEBUG nova.network.neutron [-] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.475869] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.643288] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Acquiring lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.643288] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Acquired lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.643288] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.643288] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.731061] env[61911]: INFO nova.compute.manager [-] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Took 1.04 seconds to deallocate network for instance. [ 640.733153] env[61911]: DEBUG nova.compute.claims [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.733338] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.958144] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6112df9c-e4bf-4471-9e59-3d238c0ef851 tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.959404] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.939s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.149938] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-c4a4e40a-c3d3-4d27-98a4-452f18c6408d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.149938] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 641.150139] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.150682] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.182699] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.189215] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.324554] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.693274] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.831863] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Releasing lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.832343] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 641.833068] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.833609] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a6a53d3-9505-439e-8436-15f6a8cd0377 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.845167] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bebb54-e018-4408-b2d2-54e20897e46f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.870067] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0a592ff4-3de3-41a2-9f34-a552203b66a2 could not be found. [ 641.870067] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.870067] env[61911]: INFO nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 641.870067] env[61911]: DEBUG oslo.service.loopingcall [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.872539] env[61911]: DEBUG nova.compute.manager [-] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.872665] env[61911]: DEBUG nova.network.neutron [-] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.883517] env[61911]: DEBUG nova.compute.manager [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Received event network-changed-38f28b29-a1fa-4315-9642-5604bd7c3346 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 641.883833] env[61911]: DEBUG nova.compute.manager [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Refreshing instance network info cache due to event network-changed-38f28b29-a1fa-4315-9642-5604bd7c3346. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 641.883906] env[61911]: DEBUG oslo_concurrency.lockutils [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] Acquiring lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.884065] env[61911]: DEBUG oslo_concurrency.lockutils [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] Acquired lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.884229] env[61911]: DEBUG nova.network.neutron [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Refreshing network info cache for port 38f28b29-a1fa-4315-9642-5604bd7c3346 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.945608] env[61911]: DEBUG nova.network.neutron [-] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.968719] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89f9dcf-324b-42ad-8e32-cd7ecc1fa2dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.977012] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece48e31-98eb-4f15-9d85-3e3e03bed026 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.007035] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b514253a-2b7c-43bf-89c4-dc82a6b41200 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.014689] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d675f4-5b93-4b33-8f8b-18af0767f8ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.028328] env[61911]: DEBUG nova.compute.provider_tree [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.196732] env[61911]: INFO nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: c4a4e40a-c3d3-4d27-98a4-452f18c6408d] Took 1.05 seconds to deallocate network for instance. [ 642.412632] env[61911]: DEBUG nova.network.neutron [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.453019] env[61911]: DEBUG nova.network.neutron [-] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.534394] env[61911]: DEBUG nova.scheduler.client.report [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 642.591737] env[61911]: DEBUG nova.network.neutron [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.955186] env[61911]: INFO nova.compute.manager [-] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Took 1.08 seconds to deallocate network for instance. [ 642.957879] env[61911]: DEBUG nova.compute.claims [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.958664] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.038844] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.079s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.039491] env[61911]: ERROR nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Traceback (most recent call last): [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self.driver.spawn(context, instance, image_meta, [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] vm_ref = self.build_virtual_machine(instance, [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.039491] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] for vif in network_info: [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] return self._sync_wrapper(fn, *args, **kwargs) [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self.wait() [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self[:] = self._gt.wait() [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] return self._exit_event.wait() [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] current.throw(*self._exc) [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.039803] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] result = function(*args, **kwargs) [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] return func(*args, **kwargs) [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] raise e [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] nwinfo = self.network_api.allocate_for_instance( [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] created_port_ids = self._update_ports_for_instance( [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] with excutils.save_and_reraise_exception(): [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] self.force_reraise() [ 643.040156] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] raise self.value [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] updated_port = self._update_port( [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] _ensure_no_port_binding_failure(port) [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] raise exception.PortBindingFailed(port_id=port['id']) [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] nova.exception.PortBindingFailed: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. [ 643.040463] env[61911]: ERROR nova.compute.manager [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] [ 643.040463] env[61911]: DEBUG nova.compute.utils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.041580] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.225s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.043167] env[61911]: INFO nova.compute.claims [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.046722] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Build of instance 23376481-f0eb-4ccf-b68f-1aca112a4a2d was re-scheduled: Binding failed for port 0567fcb8-b74b-405a-b78f-4d9413d5d5ae, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 643.047204] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 643.047441] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.047588] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.047747] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.095156] env[61911]: DEBUG oslo_concurrency.lockutils [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] Releasing lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.095426] env[61911]: DEBUG nova.compute.manager [req-59c6ea47-7038-4ab7-aa97-1fa9190ca9ca req-c1d35f80-1a23-4ef8-b7ba-c339faf8592e service nova] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Received event network-vif-deleted-38f28b29-a1fa-4315-9642-5604bd7c3346 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 643.266401] env[61911]: INFO nova.scheduler.client.report [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Deleted allocations for instance c4a4e40a-c3d3-4d27-98a4-452f18c6408d [ 643.578101] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.736924] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.777903] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "c4a4e40a-c3d3-4d27-98a4-452f18c6408d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.466s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.240200] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-23376481-f0eb-4ccf-b68f-1aca112a4a2d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.240200] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 644.240383] env[61911]: DEBUG nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 644.240553] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.277915] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.284829] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 644.555098] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c084d6-6696-4437-bb5d-f46eb412e014 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.562978] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09f63eb-9234-4633-bfd1-b1fa58c186b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.566378] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.566567] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 644.596542] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a42e2e-c7db-4895-92a3-d481014f65cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.605276] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f5719f-60c1-40b9-a048-3889c30768d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.620151] env[61911]: DEBUG nova.compute.provider_tree [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.790331] env[61911]: DEBUG nova.network.neutron [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.811727] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.072423] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 645.073737] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 645.073737] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Rebuilding the list of instances to heal {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 645.125918] env[61911]: DEBUG nova.scheduler.client.report [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 645.295159] env[61911]: INFO nova.compute.manager [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: 23376481-f0eb-4ccf-b68f-1aca112a4a2d] Took 1.05 seconds to deallocate network for instance. [ 645.443810] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Acquiring lock "b712ffa7-19d8-47d8-8d60-48ca52735776" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.443988] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Lock "b712ffa7-19d8-47d8-8d60-48ca52735776" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.578213] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 645.578213] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 645.578213] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 645.578213] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 645.578213] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 645.578213] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 645.605616] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-821d3f6f-e847-4ddb-ac00-ea55af302383" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.605808] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-821d3f6f-e847-4ddb-ac00-ea55af302383" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.605913] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 645.606122] env[61911]: DEBUG nova.objects.instance [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lazy-loading 'info_cache' on Instance uuid 821d3f6f-e847-4ddb-ac00-ea55af302383 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 645.628804] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.629457] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 645.631949] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.199s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.632145] env[61911]: DEBUG nova.objects.instance [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 646.137641] env[61911]: DEBUG nova.compute.utils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 646.142536] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 646.142536] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 646.217248] env[61911]: DEBUG nova.policy [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d0056b50ef7429ea460ff4c1156d132', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf7a7083777a43a5bb9fc04584dcdd06', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.338452] env[61911]: INFO nova.scheduler.client.report [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Deleted allocations for instance 23376481-f0eb-4ccf-b68f-1aca112a4a2d [ 646.630177] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.643356] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 646.650338] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3f4948a0-f3df-4e08-87ca-09f48951b871 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.651504] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.759s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.783784] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Successfully created port: 417ef611-14e7-4c8a-a5d8-f96145aecdc3 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.850263] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e44c1b50-2218-4424-9dfc-f9ee4134f14f tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "23376481-f0eb-4ccf-b68f-1aca112a4a2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.480s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.229374] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.353040] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 647.607934] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7dee2f-4746-4d97-b397-e1f04a52c86e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.616760] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1564f734-1f09-4d31-ac6f-fafa40a201b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.653502] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dacd4ef2-92a3-496e-87b2-d978c6082775 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.659157] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 647.666424] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93edbeb4-1a12-494e-a691-ebc8420c72e5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.686189] env[61911]: DEBUG nova.compute.provider_tree [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.695946] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 647.696242] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.696413] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 647.696597] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.696744] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 647.696889] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 647.697105] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 647.697272] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 647.697439] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 647.697600] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 647.697785] env[61911]: DEBUG nova.virt.hardware [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 647.698828] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b43eb6-64c4-4fd4-9eec-977cb9746b1b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.706458] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa71565-3785-471f-917f-edde19ffc77b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.735548] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-821d3f6f-e847-4ddb-ac00-ea55af302383" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.735770] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 647.735975] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.736161] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.736312] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.736460] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.736603] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.737630] env[61911]: ERROR nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. [ 647.737630] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.737630] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.737630] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.737630] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.737630] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.737630] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.737630] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.737630] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.737630] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 647.737630] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.737630] env[61911]: ERROR nova.compute.manager raise self.value [ 647.737630] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.737630] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.737630] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.737630] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.738259] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.738259] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.738259] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. [ 647.738259] env[61911]: ERROR nova.compute.manager [ 647.738259] env[61911]: Traceback (most recent call last): [ 647.738259] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.738259] env[61911]: listener.cb(fileno) [ 647.738259] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.738259] env[61911]: result = function(*args, **kwargs) [ 647.738259] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.738259] env[61911]: return func(*args, **kwargs) [ 647.738259] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.738259] env[61911]: raise e [ 647.738259] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.738259] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 647.738259] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.738259] env[61911]: created_port_ids = self._update_ports_for_instance( [ 647.738259] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.738259] env[61911]: with excutils.save_and_reraise_exception(): [ 647.738259] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.738259] env[61911]: self.force_reraise() [ 647.738259] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.738259] env[61911]: raise self.value [ 647.738259] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.738259] env[61911]: updated_port = self._update_port( [ 647.738259] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.738259] env[61911]: _ensure_no_port_binding_failure(port) [ 647.738259] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.738259] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.739079] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. [ 647.739079] env[61911]: Removing descriptor: 16 [ 647.739079] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.739079] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 647.739079] env[61911]: ERROR nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. [ 647.739079] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Traceback (most recent call last): [ 647.739079] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 647.739079] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] yield resources [ 647.739079] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 647.739079] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self.driver.spawn(context, instance, image_meta, [ 647.739079] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] vm_ref = self.build_virtual_machine(instance, [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] for vif in network_info: [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return self._sync_wrapper(fn, *args, **kwargs) [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self.wait() [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.739395] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self[:] = self._gt.wait() [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return self._exit_event.wait() [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] result = hub.switch() [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return self.greenlet.switch() [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] result = function(*args, **kwargs) [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return func(*args, **kwargs) [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] raise e [ 647.739730] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] nwinfo = self.network_api.allocate_for_instance( [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] created_port_ids = self._update_ports_for_instance( [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] with excutils.save_and_reraise_exception(): [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self.force_reraise() [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] raise self.value [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] updated_port = self._update_port( [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.740110] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] _ensure_no_port_binding_failure(port) [ 647.740451] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.740451] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] raise exception.PortBindingFailed(port_id=port['id']) [ 647.740451] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] nova.exception.PortBindingFailed: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. [ 647.740451] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] [ 647.740451] env[61911]: INFO nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Terminating instance [ 647.740955] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 647.826083] env[61911]: DEBUG nova.compute.manager [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Received event network-changed-417ef611-14e7-4c8a-a5d8-f96145aecdc3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 647.826339] env[61911]: DEBUG nova.compute.manager [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Refreshing instance network info cache due to event network-changed-417ef611-14e7-4c8a-a5d8-f96145aecdc3. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 647.826510] env[61911]: DEBUG oslo_concurrency.lockutils [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] Acquiring lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.826631] env[61911]: DEBUG oslo_concurrency.lockutils [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] Acquired lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.826793] env[61911]: DEBUG nova.network.neutron [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Refreshing network info cache for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 647.871918] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.189773] env[61911]: DEBUG nova.scheduler.client.report [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 648.247365] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.247808] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.346231] env[61911]: DEBUG nova.network.neutron [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.423476] env[61911]: DEBUG nova.network.neutron [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.701221] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.701221] env[61911]: ERROR nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. [ 648.701221] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Traceback (most recent call last): [ 648.701221] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 648.701221] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self.driver.spawn(context, instance, image_meta, [ 648.701221] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 648.701221] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.701221] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.701221] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] vm_ref = self.build_virtual_machine(instance, [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] for vif in network_info: [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return self._sync_wrapper(fn, *args, **kwargs) [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self.wait() [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self[:] = self._gt.wait() [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return self._exit_event.wait() [ 648.701633] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] result = hub.switch() [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return self.greenlet.switch() [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] result = function(*args, **kwargs) [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] return func(*args, **kwargs) [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] raise e [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] nwinfo = self.network_api.allocate_for_instance( [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.701924] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] created_port_ids = self._update_ports_for_instance( [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] with excutils.save_and_reraise_exception(): [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] self.force_reraise() [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] raise self.value [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] updated_port = self._update_port( [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] _ensure_no_port_binding_failure(port) [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.702303] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] raise exception.PortBindingFailed(port_id=port['id']) [ 648.702568] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] nova.exception.PortBindingFailed: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. [ 648.702568] env[61911]: ERROR nova.compute.manager [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] [ 648.702568] env[61911]: DEBUG nova.compute.utils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.704455] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.904s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.707157] env[61911]: INFO nova.compute.claims [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.709217] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Build of instance f6d437f9-c341-4f4b-9b68-63711ca0ed42 was re-scheduled: Binding failed for port a377842f-2c4f-4b9f-9938-a33db11ae3b1, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 648.709821] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 648.710822] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Acquiring lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.710822] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Acquired lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.710822] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.926408] env[61911]: DEBUG oslo_concurrency.lockutils [req-4a6b2663-2e70-4f96-835a-21deb109ce19 req-c566a831-5755-4ebe-92da-d75826473c2c service nova] Releasing lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.926846] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.927043] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.190681] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "aa8d4787-0638-4bc9-99aa-01678272f59a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.190814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "aa8d4787-0638-4bc9-99aa-01678272f59a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.224227] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "a77a9374-0220-43ec-bf51-34523bc8babd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.224505] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "a77a9374-0220-43ec-bf51-34523bc8babd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.240318] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.369015] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.449131] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.561836] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.873132] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Releasing lock "refresh_cache-f6d437f9-c341-4f4b-9b68-63711ca0ed42" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.873310] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 649.873435] env[61911]: DEBUG nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 649.873607] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.902372] env[61911]: DEBUG nova.compute.manager [req-e2b5f7c8-edb3-42b4-8023-34fd58822d31 req-941ca496-480d-403a-b309-c0ef454cb804 service nova] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Received event network-vif-deleted-417ef611-14e7-4c8a-a5d8-f96145aecdc3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 649.903386] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.065605] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.066051] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 650.066213] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.066870] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a8ccf2d-bb98-4b97-a97c-b6405592a888 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.076956] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95df1091-4faa-4551-ad95-e95cf4dd1265 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.112992] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bf3d8ee0-3da2-4d18-8fe5-0472118491a5 could not be found. [ 650.114560] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.114908] env[61911]: INFO nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 650.115286] env[61911]: DEBUG oslo.service.loopingcall [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.120563] env[61911]: DEBUG nova.compute.manager [-] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 650.120777] env[61911]: DEBUG nova.network.neutron [-] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.142120] env[61911]: DEBUG nova.network.neutron [-] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.190389] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c2f307-4e71-47b8-b384-d3d1c1077e80 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.199667] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3b0726-e44c-4751-812e-45e18b91889b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.240453] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c6c4f1-3c34-49ae-a11f-520b6db04fe6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.247866] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcd56c1-a22d-4dc0-b349-55740a43dbca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.262212] env[61911]: DEBUG nova.compute.provider_tree [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.405366] env[61911]: DEBUG nova.network.neutron [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.644500] env[61911]: DEBUG nova.network.neutron [-] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.765072] env[61911]: DEBUG nova.scheduler.client.report [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 650.911302] env[61911]: INFO nova.compute.manager [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] [instance: f6d437f9-c341-4f4b-9b68-63711ca0ed42] Took 1.04 seconds to deallocate network for instance. [ 651.146714] env[61911]: INFO nova.compute.manager [-] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Took 1.03 seconds to deallocate network for instance. [ 651.149112] env[61911]: DEBUG nova.compute.claims [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.149301] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.277022] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.277571] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 651.281011] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.311s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.281154] env[61911]: DEBUG nova.objects.instance [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lazy-loading 'resources' on Instance uuid 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 651.786314] env[61911]: DEBUG nova.compute.utils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 651.786450] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 651.786592] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 651.879392] env[61911]: DEBUG nova.policy [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '142ced18076b4e1d80a56bae753dce9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbcd0cd6f3cf4c20ba385b874ba34fa1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.947846] env[61911]: INFO nova.scheduler.client.report [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Deleted allocations for instance f6d437f9-c341-4f4b-9b68-63711ca0ed42 [ 652.246211] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f223bc34-f856-4481-9a78-62493e0953f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.252449] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042d6eba-df71-438e-a664-b1fa3b721126 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.293389] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 652.296839] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00bd5584-eec3-40ae-830f-588c5298dcf1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.310096] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee93854-3a87-4eaa-a6f0-b18ce6483962 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.321391] env[61911]: DEBUG nova.compute.provider_tree [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.459147] env[61911]: DEBUG oslo_concurrency.lockutils [None req-16e3233c-f128-4f53-8241-24c8e975b713 tempest-ServerActionsTestJSON-1801596596 tempest-ServerActionsTestJSON-1801596596-project-member] Lock "f6d437f9-c341-4f4b-9b68-63711ca0ed42" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.128s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.491560] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Successfully created port: 4d62c86d-8000-4f3b-a18d-7dcd62dcded1 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.824638] env[61911]: DEBUG nova.scheduler.client.report [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 652.907247] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Successfully created port: f0e6d516-b7f3-4a26-b6f7-26f7a49083a3 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.966240] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 653.305056] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 653.329787] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.049s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.333313] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.996s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.333313] env[61911]: DEBUG nova.objects.instance [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 653.340939] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 653.341278] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.341377] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 653.341558] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.341699] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 653.341850] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 653.346537] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 653.346537] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 653.346537] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 653.346736] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 653.347172] env[61911]: DEBUG nova.virt.hardware [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 653.347881] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cfcd37-517c-4252-a199-9f7150714f73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.357095] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85834ea-901f-4b72-a854-1fb58f30b3de {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.380730] env[61911]: INFO nova.scheduler.client.report [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Deleted allocations for instance 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e [ 653.489832] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.598768] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Successfully created port: a84c2bc1-9f69-4f0a-b7d3-e6c06a6437b5 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.890669] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2aa2296-f9c2-474b-8bca-8a3ecda59a77 tempest-ServerShowV254Test-1323467604 tempest-ServerShowV254Test-1323467604-project-member] Lock "7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.802s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.358209] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b1efe92a-0117-41aa-a9be-5600bf6a7be1 tempest-ServersAdmin275Test-1415112784 tempest-ServersAdmin275Test-1415112784-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.025s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.359217] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.940s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.141351] env[61911]: DEBUG nova.compute.manager [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Received event network-changed-4d62c86d-8000-4f3b-a18d-7dcd62dcded1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 655.143088] env[61911]: DEBUG nova.compute.manager [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Refreshing instance network info cache due to event network-changed-4d62c86d-8000-4f3b-a18d-7dcd62dcded1. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 655.143088] env[61911]: DEBUG oslo_concurrency.lockutils [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] Acquiring lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.143088] env[61911]: DEBUG oslo_concurrency.lockutils [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] Acquired lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.143088] env[61911]: DEBUG nova.network.neutron [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Refreshing network info cache for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.170256] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Acquiring lock "d4f17919-9cc9-4e19-88eb-38a978115975" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.170492] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Lock "d4f17919-9cc9-4e19-88eb-38a978115975" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.329929] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b238d381-b87e-412f-8290-f8cbd276eb8a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.333653] env[61911]: ERROR nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. [ 655.333653] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 655.333653] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 655.333653] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 655.333653] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.333653] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 655.333653] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.333653] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 655.333653] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.333653] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 655.333653] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.333653] env[61911]: ERROR nova.compute.manager raise self.value [ 655.333653] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.333653] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 655.333653] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.333653] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 655.334150] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.334150] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 655.334150] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. [ 655.334150] env[61911]: ERROR nova.compute.manager [ 655.334150] env[61911]: Traceback (most recent call last): [ 655.334150] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 655.334150] env[61911]: listener.cb(fileno) [ 655.334150] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.334150] env[61911]: result = function(*args, **kwargs) [ 655.334150] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.334150] env[61911]: return func(*args, **kwargs) [ 655.334150] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 655.334150] env[61911]: raise e [ 655.334150] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 655.334150] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 655.334150] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.334150] env[61911]: created_port_ids = self._update_ports_for_instance( [ 655.334150] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.334150] env[61911]: with excutils.save_and_reraise_exception(): [ 655.334150] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.334150] env[61911]: self.force_reraise() [ 655.334150] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.334150] env[61911]: raise self.value [ 655.334150] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.334150] env[61911]: updated_port = self._update_port( [ 655.334150] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.334150] env[61911]: _ensure_no_port_binding_failure(port) [ 655.334150] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.334150] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 655.334945] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. [ 655.334945] env[61911]: Removing descriptor: 16 [ 655.334945] env[61911]: ERROR nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Traceback (most recent call last): [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] yield resources [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self.driver.spawn(context, instance, image_meta, [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.334945] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] vm_ref = self.build_virtual_machine(instance, [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] for vif in network_info: [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return self._sync_wrapper(fn, *args, **kwargs) [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self.wait() [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self[:] = self._gt.wait() [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return self._exit_event.wait() [ 655.335313] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] result = hub.switch() [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return self.greenlet.switch() [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] result = function(*args, **kwargs) [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return func(*args, **kwargs) [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] raise e [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] nwinfo = self.network_api.allocate_for_instance( [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.335667] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] created_port_ids = self._update_ports_for_instance( [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] with excutils.save_and_reraise_exception(): [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self.force_reraise() [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] raise self.value [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] updated_port = self._update_port( [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] _ensure_no_port_binding_failure(port) [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.336029] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] raise exception.PortBindingFailed(port_id=port['id']) [ 655.336363] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] nova.exception.PortBindingFailed: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. [ 655.336363] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] [ 655.336363] env[61911]: INFO nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Terminating instance [ 655.341757] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd4c093-3e69-45d1-9f7b-7f2a5f413f38 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.376119] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f7d287-b410-4223-8c62-81e6566ffce9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.386029] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b1b6e6-cb20-4ae5-8641-7105b6fda836 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.408039] env[61911]: DEBUG nova.compute.provider_tree [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.695219] env[61911]: DEBUG nova.network.neutron [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.818185] env[61911]: DEBUG nova.network.neutron [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.839009] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.911303] env[61911]: DEBUG nova.scheduler.client.report [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.321068] env[61911]: DEBUG oslo_concurrency.lockutils [req-73c17cc1-c6c9-4dd4-b72a-77a7f2d0465d req-c5dfffe2-21c4-49e7-b24a-d1156ec6a1a0 service nova] Releasing lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.321574] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquired lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.321774] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.416661] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.417203] env[61911]: ERROR nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Traceback (most recent call last): [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self.driver.spawn(context, instance, image_meta, [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] vm_ref = self.build_virtual_machine(instance, [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.417203] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] for vif in network_info: [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return self._sync_wrapper(fn, *args, **kwargs) [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self.wait() [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self[:] = self._gt.wait() [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return self._exit_event.wait() [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] result = hub.switch() [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.417496] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return self.greenlet.switch() [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] result = function(*args, **kwargs) [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] return func(*args, **kwargs) [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] raise e [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] nwinfo = self.network_api.allocate_for_instance( [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] created_port_ids = self._update_ports_for_instance( [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] with excutils.save_and_reraise_exception(): [ 656.417789] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] self.force_reraise() [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] raise self.value [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] updated_port = self._update_port( [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] _ensure_no_port_binding_failure(port) [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] raise exception.PortBindingFailed(port_id=port['id']) [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] nova.exception.PortBindingFailed: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. [ 656.418169] env[61911]: ERROR nova.compute.manager [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] [ 656.418432] env[61911]: DEBUG nova.compute.utils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.419238] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.922s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.419456] env[61911]: DEBUG nova.objects.instance [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lazy-loading 'resources' on Instance uuid 90e2a5db-918a-4f66-a9c8-ef41dc4b855b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 656.421087] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Build of instance 95733825-2618-4f6a-b2eb-bdef3a7f60de was re-scheduled: Binding failed for port bdcb4666-4888-4dfc-9970-e4403d9f11cc, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 656.421570] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 656.421797] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Acquiring lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.421942] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Acquired lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.422116] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.868555] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.955829] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.962253] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.127772] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.169289] env[61911]: DEBUG nova.compute.manager [req-79c1a2e7-5188-43b0-9853-daa09b944d84 req-e14bf8ec-bf54-424b-b5e5-e4bf9895772c service nova] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Received event network-vif-deleted-4d62c86d-8000-4f3b-a18d-7dcd62dcded1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 657.288154] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42e8754-c14f-40b6-a536-b6a9737a2d79 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.295683] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aed7ef5-86c4-44c9-9141-c7f383e00537 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.327374] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadadcb0-8b95-4d21-a202-2135234ef795 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.334929] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15933f11-0ba0-4fd7-be8d-5f5e222a471f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.349439] env[61911]: DEBUG nova.compute.provider_tree [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.461765] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Releasing lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.462217] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 657.462431] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.462752] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0abfe19-b188-40a7-9cdc-53185f125a11 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.472162] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa00ddd-5d95-4d9f-a159-85f3a6b7616b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.493248] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5538f850-908d-4c89-8089-98daf1709bc1 could not be found. [ 657.493479] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.493709] env[61911]: INFO nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 657.493981] env[61911]: DEBUG oslo.service.loopingcall [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.494207] env[61911]: DEBUG nova.compute.manager [-] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 657.494309] env[61911]: DEBUG nova.network.neutron [-] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.549766] env[61911]: DEBUG nova.network.neutron [-] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.633303] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Releasing lock "refresh_cache-95733825-2618-4f6a-b2eb-bdef3a7f60de" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.633575] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 657.633763] env[61911]: DEBUG nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 657.633989] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.649602] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.854154] env[61911]: DEBUG nova.scheduler.client.report [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 658.153644] env[61911]: DEBUG nova.network.neutron [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.360152] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.363876] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.403s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.365464] env[61911]: INFO nova.compute.claims [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.405252] env[61911]: INFO nova.scheduler.client.report [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Deleted allocations for instance 90e2a5db-918a-4f66-a9c8-ef41dc4b855b [ 658.655238] env[61911]: INFO nova.compute.manager [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] [instance: 95733825-2618-4f6a-b2eb-bdef3a7f60de] Took 1.02 seconds to deallocate network for instance. [ 658.685635] env[61911]: DEBUG nova.network.neutron [-] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.913333] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3dd56dde-e713-418c-8994-44f1e96179a6 tempest-ServerDiagnosticsV248Test-749289045 tempest-ServerDiagnosticsV248Test-749289045-project-member] Lock "90e2a5db-918a-4f66-a9c8-ef41dc4b855b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.657s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.189316] env[61911]: INFO nova.compute.manager [-] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Took 1.69 seconds to deallocate network for instance. [ 659.198142] env[61911]: DEBUG nova.compute.claims [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.198142] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.692064] env[61911]: INFO nova.scheduler.client.report [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Deleted allocations for instance 95733825-2618-4f6a-b2eb-bdef3a7f60de [ 659.813343] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0e73bc-af60-4aa9-8729-502344810079 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.821399] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e47945-14b7-4bf5-a761-ee63b63f426f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.855512] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a322172f-9fd9-45c4-b94a-f1543d55ce3f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.862153] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "21723e9a-5c6d-4142-baba-25c59a5038ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.862489] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.867925] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f20c78-2ee0-4836-b506-e18023ff6e6f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.882267] env[61911]: DEBUG nova.compute.provider_tree [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.206680] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cf1955fe-148d-4971-bec0-afc111c7e76c tempest-InstanceActionsNegativeTestJSON-107135995 tempest-InstanceActionsNegativeTestJSON-107135995-project-member] Lock "95733825-2618-4f6a-b2eb-bdef3a7f60de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.789s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.388010] env[61911]: DEBUG nova.scheduler.client.report [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 660.713672] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 660.898427] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.898961] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 660.906109] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.825s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.906109] env[61911]: DEBUG nova.objects.instance [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lazy-loading 'resources' on Instance uuid 821d3f6f-e847-4ddb-ac00-ea55af302383 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 661.239468] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.411239] env[61911]: DEBUG nova.compute.utils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 661.417490] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 661.417490] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 661.518152] env[61911]: DEBUG nova.policy [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c92a35013de4848bee946186786e399', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bcf1b395cdf4ced857d542d13be9c77', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.844913] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2aaebfa-362f-496a-b4fd-96d27a09d28b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.852987] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0e24e7-c81f-4ad0-9252-d7d0ae7063b0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.883398] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034a01f2-49c0-408c-aec6-4e6cfcb5e7d1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.890590] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb6e903-e5a3-4975-8a1c-a669c9e327dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.904032] env[61911]: DEBUG nova.compute.provider_tree [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.917074] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 662.132172] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Successfully created port: 4f3ded59-aba6-4f1c-832f-7a717c954f83 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.407088] env[61911]: DEBUG nova.scheduler.client.report [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 662.912048] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.010s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.920959] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.269s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.926593] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 662.954129] env[61911]: INFO nova.scheduler.client.report [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Deleted allocations for instance 821d3f6f-e847-4ddb-ac00-ea55af302383 [ 662.970093] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 662.970418] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.970572] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 662.970757] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.971541] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 662.971921] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 662.972275] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 662.972530] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 662.974181] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 662.974181] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 662.974181] env[61911]: DEBUG nova.virt.hardware [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 662.974324] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a8f35a-c538-4c3a-babe-b9ac9fc93715 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.985107] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b4a8a4-de27-45f9-a2c2-23f8bed642a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.251111] env[61911]: DEBUG nova.compute.manager [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Received event network-changed-4f3ded59-aba6-4f1c-832f-7a717c954f83 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 663.251321] env[61911]: DEBUG nova.compute.manager [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Refreshing instance network info cache due to event network-changed-4f3ded59-aba6-4f1c-832f-7a717c954f83. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 663.251784] env[61911]: DEBUG oslo_concurrency.lockutils [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] Acquiring lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.251784] env[61911]: DEBUG oslo_concurrency.lockutils [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] Acquired lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.251936] env[61911]: DEBUG nova.network.neutron [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Refreshing network info cache for port 4f3ded59-aba6-4f1c-832f-7a717c954f83 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.466262] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d777977-8009-48fe-b052-d1af21256ded tempest-ServersAdmin275Test-1663485503 tempest-ServersAdmin275Test-1663485503-project-member] Lock "821d3f6f-e847-4ddb-ac00-ea55af302383" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.667s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.514728] env[61911]: ERROR nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. [ 663.514728] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.514728] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 663.514728] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.514728] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.514728] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.514728] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.514728] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.514728] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.514728] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 663.514728] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.514728] env[61911]: ERROR nova.compute.manager raise self.value [ 663.514728] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.514728] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.514728] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.514728] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.516117] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.516117] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.516117] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. [ 663.516117] env[61911]: ERROR nova.compute.manager [ 663.516117] env[61911]: Traceback (most recent call last): [ 663.516117] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.516117] env[61911]: listener.cb(fileno) [ 663.516117] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.516117] env[61911]: result = function(*args, **kwargs) [ 663.516117] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.516117] env[61911]: return func(*args, **kwargs) [ 663.516117] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 663.516117] env[61911]: raise e [ 663.516117] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 663.516117] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 663.516117] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.516117] env[61911]: created_port_ids = self._update_ports_for_instance( [ 663.516117] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.516117] env[61911]: with excutils.save_and_reraise_exception(): [ 663.516117] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.516117] env[61911]: self.force_reraise() [ 663.516117] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.516117] env[61911]: raise self.value [ 663.516117] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.516117] env[61911]: updated_port = self._update_port( [ 663.516117] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.516117] env[61911]: _ensure_no_port_binding_failure(port) [ 663.516117] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.516117] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.517289] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. [ 663.517289] env[61911]: Removing descriptor: 19 [ 663.517289] env[61911]: ERROR nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Traceback (most recent call last): [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] yield resources [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self.driver.spawn(context, instance, image_meta, [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.517289] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] vm_ref = self.build_virtual_machine(instance, [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] for vif in network_info: [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return self._sync_wrapper(fn, *args, **kwargs) [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self.wait() [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self[:] = self._gt.wait() [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return self._exit_event.wait() [ 663.517733] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] result = hub.switch() [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return self.greenlet.switch() [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] result = function(*args, **kwargs) [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return func(*args, **kwargs) [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] raise e [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] nwinfo = self.network_api.allocate_for_instance( [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.518569] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] created_port_ids = self._update_ports_for_instance( [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] with excutils.save_and_reraise_exception(): [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self.force_reraise() [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] raise self.value [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] updated_port = self._update_port( [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] _ensure_no_port_binding_failure(port) [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.518904] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] raise exception.PortBindingFailed(port_id=port['id']) [ 663.519212] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] nova.exception.PortBindingFailed: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. [ 663.519212] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] [ 663.519212] env[61911]: INFO nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Terminating instance [ 663.787409] env[61911]: DEBUG nova.network.neutron [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.944207] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008314ba-fc40-44ff-af36-e39cf36fb977 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.952634] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55dc921-b914-45f3-b2bf-032da125ecd6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.986326] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b99e11-3926-4f39-ac6f-64540b3396e9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.994935] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d35c30-9570-49c2-bbfd-f0740d64102f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.009727] env[61911]: DEBUG nova.compute.provider_tree [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.020464] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Acquiring lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.149352] env[61911]: DEBUG nova.network.neutron [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.516562] env[61911]: DEBUG nova.scheduler.client.report [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 664.654629] env[61911]: DEBUG oslo_concurrency.lockutils [req-b81d1d19-c60b-4eac-be6d-408c625cc590 req-3b6c2b3a-011b-4095-849c-ddc3717f22ba service nova] Releasing lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.654629] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Acquired lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.654629] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.021868] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.101s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.022556] env[61911]: ERROR nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Traceback (most recent call last): [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self.driver.spawn(context, instance, image_meta, [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] vm_ref = self.build_virtual_machine(instance, [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.022556] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] for vif in network_info: [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return self._sync_wrapper(fn, *args, **kwargs) [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self.wait() [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self[:] = self._gt.wait() [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return self._exit_event.wait() [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] result = hub.switch() [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.022987] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return self.greenlet.switch() [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] result = function(*args, **kwargs) [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] return func(*args, **kwargs) [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] raise e [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] nwinfo = self.network_api.allocate_for_instance( [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] created_port_ids = self._update_ports_for_instance( [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] with excutils.save_and_reraise_exception(): [ 665.025227] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] self.force_reraise() [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] raise self.value [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] updated_port = self._update_port( [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] _ensure_no_port_binding_failure(port) [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] raise exception.PortBindingFailed(port_id=port['id']) [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] nova.exception.PortBindingFailed: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. [ 665.025558] env[61911]: ERROR nova.compute.manager [instance: 17eec473-7082-4103-baf5-b61dc0d63281] [ 665.025814] env[61911]: DEBUG nova.compute.utils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.025814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.292s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.029900] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Build of instance 17eec473-7082-4103-baf5-b61dc0d63281 was re-scheduled: Binding failed for port edca9748-9f39-4d6c-906e-3771d1540787, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 665.030442] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 665.030711] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Acquiring lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.031135] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Acquired lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.031362] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.174645] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.316205] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.358351] env[61911]: DEBUG nova.compute.manager [req-a6fb0849-a028-415c-8eb4-b885451ea53b req-d5734722-fbbc-4047-8818-ca2b7811043d service nova] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Received event network-vif-deleted-4f3ded59-aba6-4f1c-832f-7a717c954f83 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 665.555437] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.691995] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.818822] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Releasing lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.819299] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 665.819506] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.819830] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d2486a8-8892-46dd-9800-046644709cde {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.830077] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99eae18b-b421-42de-a6b6-fe5d759a45d8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.859247] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 00a4e54a-8198-402f-a1ad-2a233480bda8 could not be found. [ 665.859561] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 665.859767] env[61911]: INFO nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 665.860044] env[61911]: DEBUG oslo.service.loopingcall [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.862752] env[61911]: DEBUG nova.compute.manager [-] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 665.862856] env[61911]: DEBUG nova.network.neutron [-] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.895168] env[61911]: DEBUG nova.network.neutron [-] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.991757] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa6c224-7dbb-4658-b306-185b9fdda961 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.999784] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d76cda8-cd0a-4116-8be7-170f68485843 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.033934] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0b3b6c-7fc6-4fd8-9c4f-39d337a0ecd4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.041547] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9640111a-cc42-44cf-8c14-30dc25b1f90e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.055197] env[61911]: DEBUG nova.compute.provider_tree [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.196869] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Releasing lock "refresh_cache-17eec473-7082-4103-baf5-b61dc0d63281" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.196869] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 666.197457] env[61911]: DEBUG nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 666.197457] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.238132] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.399575] env[61911]: DEBUG nova.network.neutron [-] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.558107] env[61911]: DEBUG nova.scheduler.client.report [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 666.741439] env[61911]: DEBUG nova.network.neutron [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.902901] env[61911]: INFO nova.compute.manager [-] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Took 1.04 seconds to deallocate network for instance. [ 666.908996] env[61911]: DEBUG nova.compute.claims [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 666.908996] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.011887] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "66472f43-537d-4eb3-8d49-d40627a8809d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.012331] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "66472f43-537d-4eb3-8d49-d40627a8809d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.065113] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.065113] env[61911]: ERROR nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. [ 667.065113] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Traceback (most recent call last): [ 667.065113] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.065113] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self.driver.spawn(context, instance, image_meta, [ 667.065113] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 667.065113] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.065113] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.065113] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] vm_ref = self.build_virtual_machine(instance, [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] for vif in network_info: [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return self._sync_wrapper(fn, *args, **kwargs) [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self.wait() [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self[:] = self._gt.wait() [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return self._exit_event.wait() [ 667.065574] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] result = hub.switch() [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return self.greenlet.switch() [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] result = function(*args, **kwargs) [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] return func(*args, **kwargs) [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] raise e [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] nwinfo = self.network_api.allocate_for_instance( [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.065879] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] created_port_ids = self._update_ports_for_instance( [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] with excutils.save_and_reraise_exception(): [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] self.force_reraise() [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] raise self.value [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] updated_port = self._update_port( [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] _ensure_no_port_binding_failure(port) [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.066199] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] raise exception.PortBindingFailed(port_id=port['id']) [ 667.066529] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] nova.exception.PortBindingFailed: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. [ 667.066529] env[61911]: ERROR nova.compute.manager [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] [ 667.066529] env[61911]: DEBUG nova.compute.utils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.066529] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.108s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.069319] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Build of instance 91e485bf-4751-4820-8f77-47bd79cc8423 was re-scheduled: Binding failed for port 208959bd-3a8b-489d-8a55-9ed7aedf7055, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 667.072252] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 667.072552] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Acquiring lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.072745] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Acquired lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.072927] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.250580] env[61911]: INFO nova.compute.manager [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] [instance: 17eec473-7082-4103-baf5-b61dc0d63281] Took 1.05 seconds to deallocate network for instance. [ 667.557011] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "267075d7-c3d8-4520-b7d9-67504097f823" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.557011] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "267075d7-c3d8-4520-b7d9-67504097f823" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.599426] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.690308] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.052882] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7c56ab-15d2-4ab9-804e-4df1c6c5cd4b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.062635] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6bf8c3-8a09-44c6-b690-eb6f88f79a69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.100976] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bd93d1-8e0c-4351-b61e-99c9d03bdc05 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.109164] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571db6a6-84ab-4f38-83d6-66f409b3e2b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.123733] env[61911]: DEBUG nova.compute.provider_tree [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.192911] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Releasing lock "refresh_cache-91e485bf-4751-4820-8f77-47bd79cc8423" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.193054] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 668.193342] env[61911]: DEBUG nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 668.193480] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.216557] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.293171] env[61911]: INFO nova.scheduler.client.report [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Deleted allocations for instance 17eec473-7082-4103-baf5-b61dc0d63281 [ 668.626784] env[61911]: DEBUG nova.scheduler.client.report [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 668.722097] env[61911]: DEBUG nova.network.neutron [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.801849] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a4e4b89b-433b-4d70-b2d5-55f584ec3401 tempest-AttachInterfacesV270Test-596598752 tempest-AttachInterfacesV270Test-596598752-project-member] Lock "17eec473-7082-4103-baf5-b61dc0d63281" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.440s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.134021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.065s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.134021] env[61911]: ERROR nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. [ 669.134021] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Traceback (most recent call last): [ 669.134021] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 669.134021] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self.driver.spawn(context, instance, image_meta, [ 669.134021] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 669.134021] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.134021] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.134021] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] vm_ref = self.build_virtual_machine(instance, [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] for vif in network_info: [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return self._sync_wrapper(fn, *args, **kwargs) [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self.wait() [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self[:] = self._gt.wait() [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return self._exit_event.wait() [ 669.134584] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] result = hub.switch() [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return self.greenlet.switch() [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] result = function(*args, **kwargs) [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] return func(*args, **kwargs) [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] raise e [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] nwinfo = self.network_api.allocate_for_instance( [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.134943] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] created_port_ids = self._update_ports_for_instance( [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] with excutils.save_and_reraise_exception(): [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] self.force_reraise() [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] raise self.value [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] updated_port = self._update_port( [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] _ensure_no_port_binding_failure(port) [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.135396] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] raise exception.PortBindingFailed(port_id=port['id']) [ 669.135728] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] nova.exception.PortBindingFailed: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. [ 669.135728] env[61911]: ERROR nova.compute.manager [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] [ 669.135728] env[61911]: DEBUG nova.compute.utils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.140034] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Build of instance 0a592ff4-3de3-41a2-9f34-a552203b66a2 was re-scheduled: Binding failed for port 38f28b29-a1fa-4315-9642-5604bd7c3346, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 669.140034] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 669.140034] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Acquiring lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.140034] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Acquired lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.140314] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.140892] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.329s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.143518] env[61911]: INFO nova.compute.claims [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 669.230619] env[61911]: INFO nova.compute.manager [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] [instance: 91e485bf-4751-4820-8f77-47bd79cc8423] Took 1.04 seconds to deallocate network for instance. [ 669.304035] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 669.679827] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.855516] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.858783] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.266612] env[61911]: INFO nova.scheduler.client.report [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Deleted allocations for instance 91e485bf-4751-4820-8f77-47bd79cc8423 [ 670.361505] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Releasing lock "refresh_cache-0a592ff4-3de3-41a2-9f34-a552203b66a2" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.361751] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 670.365016] env[61911]: DEBUG nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 670.365016] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.393511] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.695277] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bfac00-4e9b-4b3a-af48-e4d1045392f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.709359] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a295def0-bbd6-4874-864a-4a258a4f4530 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.746513] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8fa897-1369-4a6a-b085-38f046815ccc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.754257] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceef4fd9-1b8c-4a1b-96a0-43c3ae85f8f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.767510] env[61911]: DEBUG nova.compute.provider_tree [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.777770] env[61911]: DEBUG oslo_concurrency.lockutils [None req-656bbf48-e1b9-4568-94e3-a1460048aa36 tempest-AttachInterfacesUnderV243Test-318281911 tempest-AttachInterfacesUnderV243Test-318281911-project-member] Lock "91e485bf-4751-4820-8f77-47bd79cc8423" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.839s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.897373] env[61911]: DEBUG nova.network.neutron [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.270872] env[61911]: DEBUG nova.scheduler.client.report [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 671.280740] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 671.403735] env[61911]: INFO nova.compute.manager [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] [instance: 0a592ff4-3de3-41a2-9f34-a552203b66a2] Took 1.04 seconds to deallocate network for instance. [ 671.781199] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.638s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.781199] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 671.784674] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.913s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.786502] env[61911]: INFO nova.compute.claims [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.824262] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.296026] env[61911]: DEBUG nova.compute.utils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 672.298134] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 672.298489] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 672.440662] env[61911]: INFO nova.scheduler.client.report [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Deleted allocations for instance 0a592ff4-3de3-41a2-9f34-a552203b66a2 [ 672.452145] env[61911]: DEBUG nova.policy [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36d46402111342708adbc92db5637d2a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88869d5f848f42ccaadc9a80740cf2ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 672.801766] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 672.957014] env[61911]: DEBUG oslo_concurrency.lockutils [None req-837919bf-82bf-4512-9ecf-6787b6d9a51f tempest-ServerRescueTestJSONUnderV235-914363585 tempest-ServerRescueTestJSONUnderV235-914363585-project-member] Lock "0a592ff4-3de3-41a2-9f34-a552203b66a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.301s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.185181] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Successfully created port: 85e84137-d4aa-48dc-848c-4f724e1f70f7 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.250458] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b392412-7350-4e14-8068-fd184dce4f8c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.260519] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb58b8f-469b-4a6a-b44b-801a7c189218 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.295255] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca40597c-9686-4a99-a249-6e224d2baf06 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.302982] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac1e647-a493-4c14-b8a2-aa581d964678 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.326414] env[61911]: DEBUG nova.compute.provider_tree [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.459860] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 673.821575] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 673.830571] env[61911]: DEBUG nova.scheduler.client.report [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 673.861305] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 673.862043] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 673.862043] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 673.862043] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 673.864155] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 673.864322] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 673.864684] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 673.864748] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 673.864970] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 673.865136] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 673.865222] env[61911]: DEBUG nova.virt.hardware [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 673.866574] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed6cb7e-67a6-44c4-9e66-ae6b877f35b6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.874971] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ba550c-c125-4c5a-a13d-8d909f40c739 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.990678] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.338840] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.339276] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 674.345016] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 26.095s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.345016] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.345016] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 674.345016] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.193s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.346861] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced22b01-fc74-4e31-a86f-94e5b173b543 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.358649] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28450cc6-0ec1-4fa4-a429-e25a9d4c6d27 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.374380] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6969bdbf-b972-4981-bb73-690b02d039da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.383036] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ad8ce3-18b9-4942-b402-05d1f50bd57b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.419271] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181469MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 674.419541] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.856068] env[61911]: DEBUG nova.compute.utils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.857523] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 674.858815] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 674.955978] env[61911]: DEBUG nova.policy [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e40b12a3215b446782cf0dfc8e7557ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65b346cfca124fb192c8d9f5d75456a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 675.143893] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "28159e75-9fe9-44c7-b5c9-534902cecbef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.144849] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.322228] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0a6cad-58fd-4d57-a89b-f66e05245765 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.330737] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b30873a-cf3f-4eba-a0d7-a98766839d47 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.367504] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 675.369943] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160ed6d4-2c4e-45a4-b337-6322f07b6008 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.382604] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2cbfdb-4191-45a6-9054-3819ccaf0e59 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.397074] env[61911]: DEBUG nova.compute.provider_tree [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.545256] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Successfully created port: ae8c79bd-805b-46d4-bab2-81d47e27db1f {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.575250] env[61911]: DEBUG nova.compute.manager [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Received event network-changed-85e84137-d4aa-48dc-848c-4f724e1f70f7 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 675.575501] env[61911]: DEBUG nova.compute.manager [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Refreshing instance network info cache due to event network-changed-85e84137-d4aa-48dc-848c-4f724e1f70f7. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 675.575733] env[61911]: DEBUG oslo_concurrency.lockutils [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] Acquiring lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.577018] env[61911]: DEBUG oslo_concurrency.lockutils [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] Acquired lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.577260] env[61911]: DEBUG nova.network.neutron [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Refreshing network info cache for port 85e84137-d4aa-48dc-848c-4f724e1f70f7 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 675.793560] env[61911]: ERROR nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. [ 675.793560] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.793560] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 675.793560] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.793560] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.793560] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.793560] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.793560] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.793560] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.793560] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 675.793560] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.793560] env[61911]: ERROR nova.compute.manager raise self.value [ 675.793560] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.793560] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.793560] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.793560] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.794037] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.794037] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.794037] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. [ 675.794037] env[61911]: ERROR nova.compute.manager [ 675.794037] env[61911]: Traceback (most recent call last): [ 675.794037] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.794037] env[61911]: listener.cb(fileno) [ 675.794037] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.794037] env[61911]: result = function(*args, **kwargs) [ 675.794037] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.794037] env[61911]: return func(*args, **kwargs) [ 675.794037] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 675.794037] env[61911]: raise e [ 675.794037] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 675.794037] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 675.794037] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.794037] env[61911]: created_port_ids = self._update_ports_for_instance( [ 675.794037] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.794037] env[61911]: with excutils.save_and_reraise_exception(): [ 675.794037] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.794037] env[61911]: self.force_reraise() [ 675.794037] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.794037] env[61911]: raise self.value [ 675.794037] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.794037] env[61911]: updated_port = self._update_port( [ 675.794037] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.794037] env[61911]: _ensure_no_port_binding_failure(port) [ 675.794037] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.794037] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.794905] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. [ 675.794905] env[61911]: Removing descriptor: 19 [ 675.794905] env[61911]: ERROR nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Traceback (most recent call last): [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] yield resources [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self.driver.spawn(context, instance, image_meta, [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.794905] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] vm_ref = self.build_virtual_machine(instance, [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] for vif in network_info: [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return self._sync_wrapper(fn, *args, **kwargs) [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self.wait() [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self[:] = self._gt.wait() [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return self._exit_event.wait() [ 675.795251] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] result = hub.switch() [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return self.greenlet.switch() [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] result = function(*args, **kwargs) [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return func(*args, **kwargs) [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] raise e [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] nwinfo = self.network_api.allocate_for_instance( [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.795738] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] created_port_ids = self._update_ports_for_instance( [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] with excutils.save_and_reraise_exception(): [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self.force_reraise() [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] raise self.value [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] updated_port = self._update_port( [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] _ensure_no_port_binding_failure(port) [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.796126] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] raise exception.PortBindingFailed(port_id=port['id']) [ 675.796484] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] nova.exception.PortBindingFailed: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. [ 675.796484] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] [ 675.796484] env[61911]: INFO nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Terminating instance [ 675.899637] env[61911]: DEBUG nova.scheduler.client.report [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 676.104321] env[61911]: DEBUG nova.network.neutron [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.172023] env[61911]: DEBUG nova.network.neutron [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.303531] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.384338] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 676.404450] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.061s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.405134] env[61911]: ERROR nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Traceback (most recent call last): [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self.driver.spawn(context, instance, image_meta, [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] vm_ref = self.build_virtual_machine(instance, [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.405134] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] for vif in network_info: [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return self._sync_wrapper(fn, *args, **kwargs) [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self.wait() [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self[:] = self._gt.wait() [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return self._exit_event.wait() [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] result = hub.switch() [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 676.405638] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return self.greenlet.switch() [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] result = function(*args, **kwargs) [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] return func(*args, **kwargs) [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] raise e [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] nwinfo = self.network_api.allocate_for_instance( [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] created_port_ids = self._update_ports_for_instance( [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] with excutils.save_and_reraise_exception(): [ 676.405948] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] self.force_reraise() [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] raise self.value [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] updated_port = self._update_port( [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] _ensure_no_port_binding_failure(port) [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] raise exception.PortBindingFailed(port_id=port['id']) [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] nova.exception.PortBindingFailed: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. [ 676.406289] env[61911]: ERROR nova.compute.manager [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] [ 676.410183] env[61911]: DEBUG nova.compute.utils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.412171] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.919s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.414075] env[61911]: INFO nova.compute.claims [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.416892] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Build of instance bf3d8ee0-3da2-4d18-8fe5-0472118491a5 was re-scheduled: Binding failed for port 417ef611-14e7-4c8a-a5d8-f96145aecdc3, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 676.420873] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 676.421149] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.421303] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.421539] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.424517] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 676.424737] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.425015] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 676.425142] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.426086] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 676.426086] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 676.426086] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 676.426086] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 676.426086] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 676.426315] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 676.426315] env[61911]: DEBUG nova.virt.hardware [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 676.427154] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcdac88-7ae9-47c7-ab28-9638ae311340 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.437394] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e01a01-437c-45b4-9478-7baa913025b5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.674887] env[61911]: DEBUG oslo_concurrency.lockutils [req-f8fcbd17-58e9-48fc-8e57-a4941c47180b req-81c17b59-4322-4c31-9fd9-f88d12e4a3f9 service nova] Releasing lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.675807] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquired lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.675807] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.719666] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.720156] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.959836] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.099301] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.197566] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.222823] env[61911]: DEBUG nova.compute.manager [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Received event network-changed-ae8c79bd-805b-46d4-bab2-81d47e27db1f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 677.222823] env[61911]: DEBUG nova.compute.manager [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Refreshing instance network info cache due to event network-changed-ae8c79bd-805b-46d4-bab2-81d47e27db1f. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 677.223049] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] Acquiring lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.223259] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] Acquired lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.223317] env[61911]: DEBUG nova.network.neutron [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Refreshing network info cache for port ae8c79bd-805b-46d4-bab2-81d47e27db1f {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 677.268405] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.569943] env[61911]: ERROR nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. [ 677.569943] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.569943] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 677.569943] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.569943] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.569943] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.569943] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.569943] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.569943] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.569943] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 677.569943] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.569943] env[61911]: ERROR nova.compute.manager raise self.value [ 677.569943] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.569943] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.569943] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.569943] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.570662] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.570662] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.570662] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. [ 677.570662] env[61911]: ERROR nova.compute.manager [ 677.571231] env[61911]: Traceback (most recent call last): [ 677.571342] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.571342] env[61911]: listener.cb(fileno) [ 677.571428] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.571428] env[61911]: result = function(*args, **kwargs) [ 677.571499] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.571499] env[61911]: return func(*args, **kwargs) [ 677.571557] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 677.571557] env[61911]: raise e [ 677.571614] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 677.571614] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 677.571790] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.571790] env[61911]: created_port_ids = self._update_ports_for_instance( [ 677.571790] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.571790] env[61911]: with excutils.save_and_reraise_exception(): [ 677.571790] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.571790] env[61911]: self.force_reraise() [ 677.571790] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.571790] env[61911]: raise self.value [ 677.571790] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.571790] env[61911]: updated_port = self._update_port( [ 677.571790] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.571790] env[61911]: _ensure_no_port_binding_failure(port) [ 677.571790] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.571790] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.571790] env[61911]: nova.exception.PortBindingFailed: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. [ 677.571790] env[61911]: Removing descriptor: 16 [ 677.575692] env[61911]: ERROR nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Traceback (most recent call last): [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] yield resources [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self.driver.spawn(context, instance, image_meta, [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] vm_ref = self.build_virtual_machine(instance, [ 677.575692] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] for vif in network_info: [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return self._sync_wrapper(fn, *args, **kwargs) [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self.wait() [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self[:] = self._gt.wait() [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return self._exit_event.wait() [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.576246] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] result = hub.switch() [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return self.greenlet.switch() [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] result = function(*args, **kwargs) [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return func(*args, **kwargs) [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] raise e [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] nwinfo = self.network_api.allocate_for_instance( [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] created_port_ids = self._update_ports_for_instance( [ 677.576600] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] with excutils.save_and_reraise_exception(): [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self.force_reraise() [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] raise self.value [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] updated_port = self._update_port( [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] _ensure_no_port_binding_failure(port) [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] raise exception.PortBindingFailed(port_id=port['id']) [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] nova.exception.PortBindingFailed: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. [ 677.577158] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] [ 677.577554] env[61911]: INFO nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Terminating instance [ 677.611041] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "refresh_cache-bf3d8ee0-3da2-4d18-8fe5-0472118491a5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.611041] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 677.611041] env[61911]: DEBUG nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 677.611041] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.646816] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.737692] env[61911]: DEBUG nova.compute.manager [req-e4edc1fe-368d-4f0a-998a-23137fa3c4eb req-156fb18f-ff56-4603-9c23-b4907e13a674 service nova] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Received event network-vif-deleted-85e84137-d4aa-48dc-848c-4f724e1f70f7 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 677.750128] env[61911]: DEBUG nova.network.neutron [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.772210] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Releasing lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.774115] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 677.774416] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 677.774923] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92dc1f6c-9fc7-47a9-a132-425de5395b39 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.789316] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7b4694-e4a7-4ae6-b89d-6a5eda2ff977 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.820329] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f21b2aae-9d93-4d48-ba19-9905149eb5ae could not be found. [ 677.822021] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 677.822021] env[61911]: INFO nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Took 0.05 seconds to destroy the instance on the hypervisor. [ 677.822021] env[61911]: DEBUG oslo.service.loopingcall [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.822021] env[61911]: DEBUG nova.compute.manager [-] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 677.822021] env[61911]: DEBUG nova.network.neutron [-] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.853683] env[61911]: DEBUG nova.network.neutron [-] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.869105] env[61911]: DEBUG nova.network.neutron [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.986124] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33547af-b581-4ca1-b4ec-56b60f5d8d4f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.991953] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2393ae-1513-4a36-825c-b1ea85d78442 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.029756] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b88ee75-9a39-4d8a-8826-c91da64de562 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.037218] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab5a28b-b655-462f-a76f-cf3fd2954a3c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.054676] env[61911]: DEBUG nova.compute.provider_tree [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.082064] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Acquiring lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.153122] env[61911]: DEBUG nova.network.neutron [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.359699] env[61911]: DEBUG nova.network.neutron [-] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.371334] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bd09864-c051-4d06-a6fc-b5405e60664d req-adfc49d5-8cef-4cbe-a939-1f03cacc8951 service nova] Releasing lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.371982] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Acquired lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.372512] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.558070] env[61911]: DEBUG nova.scheduler.client.report [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 678.655966] env[61911]: INFO nova.compute.manager [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: bf3d8ee0-3da2-4d18-8fe5-0472118491a5] Took 1.05 seconds to deallocate network for instance. [ 678.862654] env[61911]: INFO nova.compute.manager [-] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Took 1.04 seconds to deallocate network for instance. [ 678.867193] env[61911]: DEBUG nova.compute.claims [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 678.867370] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.890701] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.974336] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.063389] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.654s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.063884] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 679.066302] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.871s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.252139] env[61911]: DEBUG nova.compute.manager [req-6b25082a-1176-41b8-a4a6-9c0a9c6547a6 req-a49d0f37-23e0-4e45-96a2-d359e0afd762 service nova] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Received event network-vif-deleted-ae8c79bd-805b-46d4-bab2-81d47e27db1f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 679.261361] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.261666] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.477234] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Releasing lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.477662] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 679.477858] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.478176] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb382d48-2ce4-45ec-9002-f1a83430ff40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.487962] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5462c461-1087-451f-b8c2-6d48666d113b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.511032] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5235a412-c833-4f7c-a884-2855e61b8966 could not be found. [ 679.511232] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.511418] env[61911]: INFO nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Took 0.03 seconds to destroy the instance on the hypervisor. [ 679.511658] env[61911]: DEBUG oslo.service.loopingcall [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.511879] env[61911]: DEBUG nova.compute.manager [-] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 679.511975] env[61911]: DEBUG nova.network.neutron [-] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.527762] env[61911]: DEBUG nova.network.neutron [-] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.571329] env[61911]: DEBUG nova.compute.utils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.576231] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 679.576482] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 679.621344] env[61911]: DEBUG nova.policy [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67aa36b6fce14243a81011df89140190', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ecdc6b08e79a4040a5cc51ef24cd6c4e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.695775] env[61911]: INFO nova.scheduler.client.report [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted allocations for instance bf3d8ee0-3da2-4d18-8fe5-0472118491a5 [ 679.922419] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Successfully created port: edbb6d57-578f-4232-ad87-d8968e20e05e {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.021027] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85dc3f0e-70c0-41f4-8bab-9e8ffdefec24 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.029024] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855effea-296c-4365-9e2c-cbab00993715 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.032162] env[61911]: DEBUG nova.network.neutron [-] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.060542] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09e3476-0e09-44e8-a470-c4063e074a08 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.071467] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d702e7-f183-4eb2-9d0a-911505b8cbd5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.086090] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 680.088867] env[61911]: DEBUG nova.compute.provider_tree [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.207585] env[61911]: DEBUG oslo_concurrency.lockutils [None req-53877aa3-699b-4e69-8f88-ab57bd34fe6a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "bf3d8ee0-3da2-4d18-8fe5-0472118491a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.505s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.560286] env[61911]: INFO nova.compute.manager [-] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Took 1.05 seconds to deallocate network for instance. [ 680.562514] env[61911]: DEBUG nova.compute.claims [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.562712] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.594691] env[61911]: DEBUG nova.scheduler.client.report [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 680.711236] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 680.769898] env[61911]: ERROR nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. [ 680.769898] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.769898] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.769898] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.769898] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.769898] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.769898] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.769898] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.769898] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.769898] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 680.769898] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.769898] env[61911]: ERROR nova.compute.manager raise self.value [ 680.769898] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.769898] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.769898] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.769898] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.770320] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.770320] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.770320] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. [ 680.770320] env[61911]: ERROR nova.compute.manager [ 680.770320] env[61911]: Traceback (most recent call last): [ 680.770320] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.770320] env[61911]: listener.cb(fileno) [ 680.770320] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.770320] env[61911]: result = function(*args, **kwargs) [ 680.770320] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.770320] env[61911]: return func(*args, **kwargs) [ 680.770320] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 680.770320] env[61911]: raise e [ 680.770320] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.770320] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 680.770320] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.770320] env[61911]: created_port_ids = self._update_ports_for_instance( [ 680.770320] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.770320] env[61911]: with excutils.save_and_reraise_exception(): [ 680.770320] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.770320] env[61911]: self.force_reraise() [ 680.770320] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.770320] env[61911]: raise self.value [ 680.770320] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.770320] env[61911]: updated_port = self._update_port( [ 680.770320] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.770320] env[61911]: _ensure_no_port_binding_failure(port) [ 680.770320] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.770320] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.771081] env[61911]: nova.exception.PortBindingFailed: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. [ 680.771081] env[61911]: Removing descriptor: 16 [ 681.099783] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 681.102718] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.103326] env[61911]: ERROR nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Traceback (most recent call last): [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self.driver.spawn(context, instance, image_meta, [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] vm_ref = self.build_virtual_machine(instance, [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.103326] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] for vif in network_info: [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return self._sync_wrapper(fn, *args, **kwargs) [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self.wait() [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self[:] = self._gt.wait() [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return self._exit_event.wait() [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] result = hub.switch() [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.103633] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return self.greenlet.switch() [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] result = function(*args, **kwargs) [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] return func(*args, **kwargs) [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] raise e [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] nwinfo = self.network_api.allocate_for_instance( [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] created_port_ids = self._update_ports_for_instance( [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] with excutils.save_and_reraise_exception(): [ 681.103985] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] self.force_reraise() [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] raise self.value [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] updated_port = self._update_port( [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] _ensure_no_port_binding_failure(port) [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] raise exception.PortBindingFailed(port_id=port['id']) [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] nova.exception.PortBindingFailed: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. [ 681.104354] env[61911]: ERROR nova.compute.manager [instance: 5538f850-908d-4c89-8089-98daf1709bc1] [ 681.104657] env[61911]: DEBUG nova.compute.utils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.105628] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.866s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.107113] env[61911]: INFO nova.compute.claims [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.110317] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Build of instance 5538f850-908d-4c89-8089-98daf1709bc1 was re-scheduled: Binding failed for port 4d62c86d-8000-4f3b-a18d-7dcd62dcded1, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 681.110980] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 681.110980] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.111125] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquired lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.111343] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.123125] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 681.123347] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.123538] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 681.123736] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.123884] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 681.124044] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 681.124252] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 681.124411] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 681.124578] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 681.124740] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 681.124913] env[61911]: DEBUG nova.virt.hardware [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 681.125954] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5616980-6f52-457a-9632-716d84ff6e62 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.133797] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30611610-b0b9-4cc6-9c55-8e96e78e63c3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.147686] env[61911]: ERROR nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Traceback (most recent call last): [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] yield resources [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self.driver.spawn(context, instance, image_meta, [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] vm_ref = self.build_virtual_machine(instance, [ 681.147686] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] for vif in network_info: [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] return self._sync_wrapper(fn, *args, **kwargs) [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self.wait() [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self[:] = self._gt.wait() [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] return self._exit_event.wait() [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.148178] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] current.throw(*self._exc) [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] result = function(*args, **kwargs) [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] return func(*args, **kwargs) [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] raise e [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] nwinfo = self.network_api.allocate_for_instance( [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] created_port_ids = self._update_ports_for_instance( [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] with excutils.save_and_reraise_exception(): [ 681.148786] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self.force_reraise() [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] raise self.value [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] updated_port = self._update_port( [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] _ensure_no_port_binding_failure(port) [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] raise exception.PortBindingFailed(port_id=port['id']) [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] nova.exception.PortBindingFailed: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. [ 681.149402] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] [ 681.149402] env[61911]: INFO nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Terminating instance [ 681.236624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.242154] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.242297] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.322839] env[61911]: DEBUG nova.compute.manager [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Received event network-changed-edbb6d57-578f-4232-ad87-d8968e20e05e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 681.323220] env[61911]: DEBUG nova.compute.manager [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Refreshing instance network info cache due to event network-changed-edbb6d57-578f-4232-ad87-d8968e20e05e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 681.323492] env[61911]: DEBUG oslo_concurrency.lockutils [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] Acquiring lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.323748] env[61911]: DEBUG oslo_concurrency.lockutils [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] Acquired lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.323988] env[61911]: DEBUG nova.network.neutron [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Refreshing network info cache for port edbb6d57-578f-4232-ad87-d8968e20e05e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.629844] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.652589] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Acquiring lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.708751] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.842266] env[61911]: DEBUG nova.network.neutron [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.914696] env[61911]: DEBUG nova.network.neutron [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.211974] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Releasing lock "refresh_cache-5538f850-908d-4c89-8089-98daf1709bc1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.211974] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 682.212308] env[61911]: DEBUG nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 682.212592] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.228885] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.417692] env[61911]: DEBUG oslo_concurrency.lockutils [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] Releasing lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.417966] env[61911]: DEBUG nova.compute.manager [req-7a84e790-ede0-4b55-96f5-93413b85940a req-726d67b8-2399-45a4-8385-bfaebb56c50f service nova] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Received event network-vif-deleted-edbb6d57-578f-4232-ad87-d8968e20e05e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 682.418336] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Acquired lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.418521] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.434574] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b93e880-bcdb-4e19-939c-698bf6af4e09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.442821] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e8d397-de7c-402c-a688-cb3912cbbc03 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.472731] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f7fcf8-87f1-4754-b729-7d62fbb34619 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.479822] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfde0000-5f2a-493f-bfdd-9e08af98da2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.492549] env[61911]: DEBUG nova.compute.provider_tree [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.732725] env[61911]: DEBUG nova.network.neutron [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.938147] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.995864] env[61911]: DEBUG nova.scheduler.client.report [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 683.010455] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.235315] env[61911]: INFO nova.compute.manager [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 5538f850-908d-4c89-8089-98daf1709bc1] Took 1.02 seconds to deallocate network for instance. [ 683.500912] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.501457] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 683.504192] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.598s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.512409] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Releasing lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.512793] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 683.513018] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.513300] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de325279-56d8-4701-aff4-7013b4034f91 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.522728] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1900bd-7bae-4d8d-963f-cc569e8705b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.545463] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5002bdba-3519-4caa-8486-c2249bd9abef could not be found. [ 683.545697] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.545883] env[61911]: INFO nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Took 0.03 seconds to destroy the instance on the hypervisor. [ 683.546133] env[61911]: DEBUG oslo.service.loopingcall [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.546574] env[61911]: DEBUG nova.compute.manager [-] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 683.546694] env[61911]: DEBUG nova.network.neutron [-] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.561231] env[61911]: DEBUG nova.network.neutron [-] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.013192] env[61911]: DEBUG nova.compute.utils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.014749] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 684.014893] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 684.053923] env[61911]: DEBUG nova.policy [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ee96b52414148b7bff1d067595848b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f40f512384d4fedb75bfc53f3195243', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.063853] env[61911]: DEBUG nova.network.neutron [-] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.264283] env[61911]: INFO nova.scheduler.client.report [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Deleted allocations for instance 5538f850-908d-4c89-8089-98daf1709bc1 [ 684.392948] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Successfully created port: 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 684.415175] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e532325b-22e7-4ccc-9a16-195e1dd7376b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.422911] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4817f3f0-ea9b-4ca1-83f8-d3100f105b51 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.452917] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4deec323-36f4-4f66-9246-a62a6e4ca696 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.459957] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a634334b-5f0a-4c34-b903-c32c41835064 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.472868] env[61911]: DEBUG nova.compute.provider_tree [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.517953] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 684.567691] env[61911]: INFO nova.compute.manager [-] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Took 1.02 seconds to deallocate network for instance. [ 684.571360] env[61911]: DEBUG nova.compute.claims [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.571360] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.774574] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eeda152-5fb8-4eb7-91ba-030257719b3c tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "5538f850-908d-4c89-8089-98daf1709bc1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.281s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.976109] env[61911]: DEBUG nova.scheduler.client.report [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 685.281023] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 685.315758] env[61911]: DEBUG nova.compute.manager [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Received event network-changed-8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 685.315758] env[61911]: DEBUG nova.compute.manager [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Refreshing instance network info cache due to event network-changed-8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 685.315758] env[61911]: DEBUG oslo_concurrency.lockutils [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] Acquiring lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.316874] env[61911]: DEBUG oslo_concurrency.lockutils [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] Acquired lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.316874] env[61911]: DEBUG nova.network.neutron [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Refreshing network info cache for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 685.353143] env[61911]: ERROR nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. [ 685.353143] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 685.353143] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 685.353143] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 685.353143] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.353143] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 685.353143] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.353143] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 685.353143] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.353143] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 685.353143] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.353143] env[61911]: ERROR nova.compute.manager raise self.value [ 685.353143] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.353143] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 685.353143] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.353143] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 685.353613] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.353613] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 685.353613] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. [ 685.353613] env[61911]: ERROR nova.compute.manager [ 685.353613] env[61911]: Traceback (most recent call last): [ 685.353613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 685.353613] env[61911]: listener.cb(fileno) [ 685.353613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.353613] env[61911]: result = function(*args, **kwargs) [ 685.353613] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.353613] env[61911]: return func(*args, **kwargs) [ 685.353613] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 685.353613] env[61911]: raise e [ 685.353613] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 685.353613] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 685.353613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.353613] env[61911]: created_port_ids = self._update_ports_for_instance( [ 685.353613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.353613] env[61911]: with excutils.save_and_reraise_exception(): [ 685.353613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.353613] env[61911]: self.force_reraise() [ 685.353613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.353613] env[61911]: raise self.value [ 685.353613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.353613] env[61911]: updated_port = self._update_port( [ 685.353613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.353613] env[61911]: _ensure_no_port_binding_failure(port) [ 685.353613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.353613] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 685.354391] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. [ 685.354391] env[61911]: Removing descriptor: 16 [ 685.482608] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.483209] env[61911]: ERROR nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Traceback (most recent call last): [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self.driver.spawn(context, instance, image_meta, [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] vm_ref = self.build_virtual_machine(instance, [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.483209] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] for vif in network_info: [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return self._sync_wrapper(fn, *args, **kwargs) [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self.wait() [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self[:] = self._gt.wait() [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return self._exit_event.wait() [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] result = hub.switch() [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.483678] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return self.greenlet.switch() [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] result = function(*args, **kwargs) [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] return func(*args, **kwargs) [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] raise e [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] nwinfo = self.network_api.allocate_for_instance( [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] created_port_ids = self._update_ports_for_instance( [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] with excutils.save_and_reraise_exception(): [ 685.484020] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] self.force_reraise() [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] raise self.value [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] updated_port = self._update_port( [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] _ensure_no_port_binding_failure(port) [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] raise exception.PortBindingFailed(port_id=port['id']) [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] nova.exception.PortBindingFailed: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. [ 685.484383] env[61911]: ERROR nova.compute.manager [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] [ 685.484704] env[61911]: DEBUG nova.compute.utils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.485188] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.630s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.486710] env[61911]: INFO nova.compute.claims [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.489459] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Build of instance 00a4e54a-8198-402f-a1ad-2a233480bda8 was re-scheduled: Binding failed for port 4f3ded59-aba6-4f1c-832f-7a717c954f83, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 685.489886] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 685.490280] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Acquiring lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.490439] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Acquired lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.490675] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.528341] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 685.567868] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 685.569182] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.573018] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 685.573018] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.573018] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 685.573018] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 685.573018] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 685.573267] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 685.573267] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 685.573267] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 685.573267] env[61911]: DEBUG nova.virt.hardware [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 685.573267] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d1aa75-8b1d-403e-86d2-f991971a7de5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.582525] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7084bb-fcab-4ebe-bbcf-6f126581c797 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.598785] env[61911]: ERROR nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Traceback (most recent call last): [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] yield resources [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self.driver.spawn(context, instance, image_meta, [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] vm_ref = self.build_virtual_machine(instance, [ 685.598785] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] for vif in network_info: [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] return self._sync_wrapper(fn, *args, **kwargs) [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self.wait() [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self[:] = self._gt.wait() [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] return self._exit_event.wait() [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 685.600362] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] current.throw(*self._exc) [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] result = function(*args, **kwargs) [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] return func(*args, **kwargs) [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] raise e [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] nwinfo = self.network_api.allocate_for_instance( [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] created_port_ids = self._update_ports_for_instance( [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] with excutils.save_and_reraise_exception(): [ 685.600944] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self.force_reraise() [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] raise self.value [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] updated_port = self._update_port( [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] _ensure_no_port_binding_failure(port) [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] raise exception.PortBindingFailed(port_id=port['id']) [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] nova.exception.PortBindingFailed: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. [ 685.603986] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] [ 685.603986] env[61911]: INFO nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Terminating instance [ 685.807937] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.836161] env[61911]: DEBUG nova.network.neutron [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.911118] env[61911]: DEBUG nova.network.neutron [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.011165] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.078610] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.105409] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.416408] env[61911]: DEBUG oslo_concurrency.lockutils [req-7d9b828e-f4fb-4020-9ef4-b5abfbc41692 req-d2774edc-2921-4587-945f-aa5882a24efd service nova] Releasing lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.416408] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.416408] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.582654] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Releasing lock "refresh_cache-00a4e54a-8198-402f-a1ad-2a233480bda8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.582916] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 686.583112] env[61911]: DEBUG nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 686.583284] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.598738] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.907632] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4586de-134f-4d2f-8471-319e8a8e1a43 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.915368] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814cc353-952b-4076-a8cf-869fc7232a5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.948100] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.950294] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fde7cd-312b-4ac1-9ae3-06b20134247c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.958065] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5718cb48-e42b-4a85-b7b1-0697696a4a69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.970822] env[61911]: DEBUG nova.compute.provider_tree [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.027066] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.101145] env[61911]: DEBUG nova.network.neutron [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.364310] env[61911]: DEBUG nova.compute.manager [req-9e408104-3b88-489a-a9cc-4fc857257071 req-6bd85320-b316-4612-a083-c60022ad21cd service nova] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Received event network-vif-deleted-8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 687.475070] env[61911]: DEBUG nova.scheduler.client.report [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.531317] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.531317] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 687.531557] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.532700] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ecc3c2b-3862-43ca-ab18-349041264ae5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.540504] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b409e2-9a96-4146-9ba4-cbe0e221028b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.561908] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61c4a92c-296a-4c0e-aa9a-32b20a837b6d could not be found. [ 687.562135] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 687.562320] env[61911]: INFO nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 687.562559] env[61911]: DEBUG oslo.service.loopingcall [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.562766] env[61911]: DEBUG nova.compute.manager [-] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 687.562861] env[61911]: DEBUG nova.network.neutron [-] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.578518] env[61911]: DEBUG nova.network.neutron [-] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.604839] env[61911]: INFO nova.compute.manager [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] [instance: 00a4e54a-8198-402f-a1ad-2a233480bda8] Took 1.02 seconds to deallocate network for instance. [ 687.810821] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "951e57ba-da09-426a-b500-e5459d346f64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.810821] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "951e57ba-da09-426a-b500-e5459d346f64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.980388] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.981339] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 687.983759] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.160s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.985298] env[61911]: INFO nova.compute.claims [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.082013] env[61911]: DEBUG nova.network.neutron [-] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.491334] env[61911]: DEBUG nova.compute.utils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.493510] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 688.493680] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 688.547528] env[61911]: DEBUG nova.policy [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c7f2fa929584f6589d3f37ed233b4e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f89681e87fd541beb1741b9d231a0e19', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.584265] env[61911]: INFO nova.compute.manager [-] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Took 1.02 seconds to deallocate network for instance. [ 688.586688] env[61911]: DEBUG nova.compute.claims [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 688.586853] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.637541] env[61911]: INFO nova.scheduler.client.report [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Deleted allocations for instance 00a4e54a-8198-402f-a1ad-2a233480bda8 [ 688.832331] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Successfully created port: c6e44c9b-bfe8-4463-8444-6ebae2c90be6 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.994484] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 689.145935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3d9e7771-0318-4da2-a2dd-566357d0ed78 tempest-ServerActionsTestOtherB-1235264950 tempest-ServerActionsTestOtherB-1235264950-project-member] Lock "00a4e54a-8198-402f-a1ad-2a233480bda8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.203s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.329575] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf66d2cb-ac52-423e-9b92-b2d83d9c12b5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.340018] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c09c87-0471-4b0d-89de-2b64041a1f9d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.378858] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f273da21-27e5-4775-a244-b4dbce8fdfbf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.386637] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08941b88-b450-4ce3-9a5c-192687b776e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.401195] env[61911]: DEBUG nova.compute.provider_tree [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.504740] env[61911]: INFO nova.virt.block_device [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Booting with volume 9d941863-3cff-4a05-a65c-29550cd7a6c6 at /dev/sda [ 689.524909] env[61911]: DEBUG nova.compute.manager [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Received event network-changed-c6e44c9b-bfe8-4463-8444-6ebae2c90be6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 689.524909] env[61911]: DEBUG nova.compute.manager [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Refreshing instance network info cache due to event network-changed-c6e44c9b-bfe8-4463-8444-6ebae2c90be6. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 689.528019] env[61911]: DEBUG oslo_concurrency.lockutils [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] Acquiring lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.528019] env[61911]: DEBUG oslo_concurrency.lockutils [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] Acquired lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.528019] env[61911]: DEBUG nova.network.neutron [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Refreshing network info cache for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 689.569466] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03c284f7-9a2b-4742-bec7-d4e022937435 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.578978] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8150d96-0cf9-4208-8323-63dccd871cc4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.601311] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-600ca121-6ffc-4ad8-9761-c06d1fe31d6c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.609819] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ee49fc-b096-4089-917b-0a10c63366c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.632933] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7da6e44-44c0-4cae-962e-ca6bde65ca41 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.639480] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe88e1d7-aee4-4195-bd87-1058fb184099 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.647275] env[61911]: ERROR nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. [ 689.647275] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.647275] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.647275] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.647275] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.647275] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.647275] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.647275] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.647275] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.647275] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 689.647275] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.647275] env[61911]: ERROR nova.compute.manager raise self.value [ 689.647275] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.647275] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.647275] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.647275] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.647762] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.647762] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.647762] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. [ 689.647762] env[61911]: ERROR nova.compute.manager [ 689.647762] env[61911]: Traceback (most recent call last): [ 689.647762] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.647762] env[61911]: listener.cb(fileno) [ 689.647762] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.647762] env[61911]: result = function(*args, **kwargs) [ 689.647762] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.647762] env[61911]: return func(*args, **kwargs) [ 689.647762] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 689.647762] env[61911]: raise e [ 689.647762] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.647762] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 689.647762] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.647762] env[61911]: created_port_ids = self._update_ports_for_instance( [ 689.647762] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.647762] env[61911]: with excutils.save_and_reraise_exception(): [ 689.647762] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.647762] env[61911]: self.force_reraise() [ 689.647762] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.647762] env[61911]: raise self.value [ 689.647762] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.647762] env[61911]: updated_port = self._update_port( [ 689.647762] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.647762] env[61911]: _ensure_no_port_binding_failure(port) [ 689.647762] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.647762] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.648459] env[61911]: nova.exception.PortBindingFailed: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. [ 689.648459] env[61911]: Removing descriptor: 16 [ 689.648459] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 689.654148] env[61911]: DEBUG nova.virt.block_device [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Updating existing volume attachment record: d704e373-75c9-4336-928b-4a10f56660f4 {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 689.904688] env[61911]: DEBUG nova.scheduler.client.report [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 690.047129] env[61911]: DEBUG nova.network.neutron [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.119871] env[61911]: DEBUG nova.network.neutron [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.173596] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.409635] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.410059] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 690.412824] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.422s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.414206] env[61911]: INFO nova.compute.claims [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.623764] env[61911]: DEBUG oslo_concurrency.lockutils [req-4cfa37ed-5904-4d23-a678-07b8d3ec9126 req-07349467-3657-4895-9bc8-03205f0a265f service nova] Releasing lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.923833] env[61911]: DEBUG nova.compute.utils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.925568] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 690.925568] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 690.975347] env[61911]: DEBUG nova.policy [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f28191f8c1fe45d9923d80847ebd052b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a2306e99fe94c7389af7aa8f62ea57c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.286925] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Successfully created port: b4c68e4c-f488-40d7-be84-76f26f71daa3 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.429687] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 691.546593] env[61911]: DEBUG nova.compute.manager [req-25288a29-595a-4e92-8f87-394bd013c9ed req-26978b25-01e0-4483-a6fb-f29009a3e25f service nova] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Received event network-vif-deleted-c6e44c9b-bfe8-4463-8444-6ebae2c90be6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 691.766751] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 691.767500] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 691.767500] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.767683] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 691.767818] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.767958] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 691.768128] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 691.768318] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 691.768473] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 691.768636] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 691.768850] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 691.769169] env[61911]: DEBUG nova.virt.hardware [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 691.771314] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae9f997-4102-4027-a892-7960afa495d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.788069] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80843b7-aff2-4275-a26f-4e5b4854d121 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.806546] env[61911]: ERROR nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Traceback (most recent call last): [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] yield resources [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self.driver.spawn(context, instance, image_meta, [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] vm_ref = self.build_virtual_machine(instance, [ 691.806546] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] for vif in network_info: [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] return self._sync_wrapper(fn, *args, **kwargs) [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self.wait() [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self[:] = self._gt.wait() [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] return self._exit_event.wait() [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 691.807345] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] current.throw(*self._exc) [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] result = function(*args, **kwargs) [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] return func(*args, **kwargs) [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] raise e [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] nwinfo = self.network_api.allocate_for_instance( [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] created_port_ids = self._update_ports_for_instance( [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] with excutils.save_and_reraise_exception(): [ 691.809308] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self.force_reraise() [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] raise self.value [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] updated_port = self._update_port( [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] _ensure_no_port_binding_failure(port) [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] raise exception.PortBindingFailed(port_id=port['id']) [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] nova.exception.PortBindingFailed: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. [ 691.809890] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] [ 691.809890] env[61911]: INFO nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Terminating instance [ 691.872249] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e20d62c-0398-4562-b5c9-aa5ba2c50b59 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.878650] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027df00e-ea09-40bf-a9d2-3868fcbfec93 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.911569] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984c3986-b518-4222-abe8-867de288a6f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.923803] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872e53bb-9267-4424-871e-31c055158651 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.937494] env[61911]: INFO nova.virt.block_device [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Booting with volume ad4019f9-2de5-4686-812b-daa4fe87544b at /dev/sda [ 691.939296] env[61911]: DEBUG nova.compute.provider_tree [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.990316] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab7ff738-f667-4122-88e7-150655579ac6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.002214] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc1d8e2-79a5-4c7d-841a-f2751758b359 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.023569] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7550c5be-d5ef-4c13-83f1-d017febd364e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.033266] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836ab60c-9173-4bc9-bc8e-3c5e037e1d40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.054343] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8027c6c9-2844-4b77-87b6-8b61ab5cc05d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.062467] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ec3ff5-e844-4ad9-8e17-6f077775fd84 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.077558] env[61911]: DEBUG nova.virt.block_device [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Updating existing volume attachment record: 78d99f69-d28e-4095-b153-5c7d25beec60 {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 692.317557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Acquiring lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.317722] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Acquired lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.317902] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.445424] env[61911]: DEBUG nova.scheduler.client.report [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 692.574142] env[61911]: ERROR nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. [ 692.574142] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.574142] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 692.574142] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.574142] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.574142] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.574142] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.574142] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.574142] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.574142] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 692.574142] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.574142] env[61911]: ERROR nova.compute.manager raise self.value [ 692.574142] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.574142] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.574142] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.574142] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.574622] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.574622] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.574622] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. [ 692.574622] env[61911]: ERROR nova.compute.manager [ 692.574622] env[61911]: Traceback (most recent call last): [ 692.574622] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.574622] env[61911]: listener.cb(fileno) [ 692.574622] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.574622] env[61911]: result = function(*args, **kwargs) [ 692.574622] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.574622] env[61911]: return func(*args, **kwargs) [ 692.574622] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 692.574622] env[61911]: raise e [ 692.574622] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 692.574622] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 692.574622] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.574622] env[61911]: created_port_ids = self._update_ports_for_instance( [ 692.574622] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.574622] env[61911]: with excutils.save_and_reraise_exception(): [ 692.574622] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.574622] env[61911]: self.force_reraise() [ 692.574622] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.574622] env[61911]: raise self.value [ 692.574622] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.574622] env[61911]: updated_port = self._update_port( [ 692.574622] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.574622] env[61911]: _ensure_no_port_binding_failure(port) [ 692.574622] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.574622] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.575391] env[61911]: nova.exception.PortBindingFailed: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. [ 692.575391] env[61911]: Removing descriptor: 16 [ 692.839345] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.923903] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.950321] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.950903] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 692.953439] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.534s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.427476] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Releasing lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.428072] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 693.428409] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de3229a1-72a5-403e-b5c2-81cafab38149 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.437638] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ee7848-40e7-4567-b716-1a2c1642b88d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.458915] env[61911]: DEBUG nova.compute.utils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.464795] env[61911]: WARNING nova.virt.vmwareapi.driver [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 08e38cae-cf76-4367-ac66-c5ecd25c2763 could not be found. [ 693.465154] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.466014] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 693.466330] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 693.468394] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d2012b3-43cf-47c0-8afb-8eee32e39721 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.476554] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e9b8bf-373f-4266-8e36-d18cfd0a5edf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.502055] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 08e38cae-cf76-4367-ac66-c5ecd25c2763 could not be found. [ 693.502295] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.502487] env[61911]: INFO nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Took 0.07 seconds to destroy the instance on the hypervisor. [ 693.502723] env[61911]: DEBUG oslo.service.loopingcall [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.502944] env[61911]: DEBUG nova.compute.manager [-] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 693.503134] env[61911]: DEBUG nova.network.neutron [-] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.519849] env[61911]: DEBUG nova.network.neutron [-] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.548805] env[61911]: DEBUG nova.policy [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96721947abba4d33b8838e88fa88c387', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7d9df8368db43609a412e2ffed933d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.574308] env[61911]: DEBUG nova.compute.manager [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Received event network-changed-b4c68e4c-f488-40d7-be84-76f26f71daa3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 693.574533] env[61911]: DEBUG nova.compute.manager [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Refreshing instance network info cache due to event network-changed-b4c68e4c-f488-40d7-be84-76f26f71daa3. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 693.574748] env[61911]: DEBUG oslo_concurrency.lockutils [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] Acquiring lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.574972] env[61911]: DEBUG oslo_concurrency.lockutils [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] Acquired lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.575339] env[61911]: DEBUG nova.network.neutron [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Refreshing network info cache for port b4c68e4c-f488-40d7-be84-76f26f71daa3 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 693.970183] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 693.993337] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance f21b2aae-9d93-4d48-ba19-9905149eb5ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 693.993509] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5235a412-c833-4f7c-a884-2855e61b8966 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 693.993636] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5002bdba-3519-4caa-8486-c2249bd9abef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 693.993753] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 61c4a92c-296a-4c0e-aa9a-32b20a837b6d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 693.993870] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 08e38cae-cf76-4367-ac66-c5ecd25c2763 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 693.993981] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance c24159b8-8974-499e-8237-26f16e3a198b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 693.994106] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance fe6c9377-f9e6-43d2-afa7-2323bf60a48d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 694.022052] env[61911]: DEBUG nova.network.neutron [-] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.095451] env[61911]: DEBUG nova.network.neutron [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.119321] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Successfully created port: 038e0939-f5ab-4535-93fb-4c4cdb1bed45 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.194446] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 694.198164] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 694.199324] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.199612] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 694.199967] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.200299] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 694.200554] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 694.204324] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 694.204324] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 694.204324] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 694.204324] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 694.204324] env[61911]: DEBUG nova.virt.hardware [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 694.204493] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175e9f13-a9a7-4bb2-ab24-d11b508a7b2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.211564] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de083e5-19cd-4276-a921-8a12c74516f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.227944] env[61911]: ERROR nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] Traceback (most recent call last): [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] yield resources [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self.driver.spawn(context, instance, image_meta, [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] vm_ref = self.build_virtual_machine(instance, [ 694.227944] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] for vif in network_info: [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] return self._sync_wrapper(fn, *args, **kwargs) [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self.wait() [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self[:] = self._gt.wait() [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] return self._exit_event.wait() [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.228370] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] current.throw(*self._exc) [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] result = function(*args, **kwargs) [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] return func(*args, **kwargs) [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] raise e [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] nwinfo = self.network_api.allocate_for_instance( [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] created_port_ids = self._update_ports_for_instance( [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] with excutils.save_and_reraise_exception(): [ 694.228791] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self.force_reraise() [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] raise self.value [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] updated_port = self._update_port( [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] _ensure_no_port_binding_failure(port) [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] raise exception.PortBindingFailed(port_id=port['id']) [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] nova.exception.PortBindingFailed: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. [ 694.229223] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] [ 694.229965] env[61911]: INFO nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Terminating instance [ 694.255633] env[61911]: DEBUG nova.network.neutron [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.497740] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance e5f4224f-0002-4b75-9f69-d1ac50d955ae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.524616] env[61911]: INFO nova.compute.manager [-] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Took 1.02 seconds to deallocate network for instance. [ 694.735188] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Acquiring lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.757968] env[61911]: DEBUG oslo_concurrency.lockutils [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] Releasing lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.758271] env[61911]: DEBUG nova.compute.manager [req-8fb2b515-2f8a-46f6-a4e6-071dd0a0d7d7 req-fed3e058-e9cb-4b52-b85c-9f3200ca2389 service nova] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Received event network-vif-deleted-b4c68e4c-f488-40d7-be84-76f26f71daa3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 694.758646] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Acquired lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.758813] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.980383] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 695.003520] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.007468] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:55:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b54539c4-8973-4fe9-8a6e-f448115d419f',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-843943453',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 695.007809] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.008037] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 695.008243] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.008394] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 695.008545] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 695.008960] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 695.009168] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 695.009342] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 695.009599] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 695.010027] env[61911]: DEBUG nova.virt.hardware [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 695.012309] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1419b6e4-c575-47e1-a3d6-1641373e05a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.019901] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832c29a1-6d90-4a7b-a4c0-6f8dd3f32b02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.087558] env[61911]: INFO nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Took 0.56 seconds to detach 1 volumes for instance. [ 695.089682] env[61911]: DEBUG nova.compute.claims [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.089859] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.277665] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.489655] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.516265] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance a09bc118-c3f8-4943-89a1-8a0c59d528c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.627386] env[61911]: ERROR nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. [ 695.627386] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.627386] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.627386] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.627386] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.627386] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.627386] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.627386] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.627386] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.627386] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 695.627386] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.627386] env[61911]: ERROR nova.compute.manager raise self.value [ 695.627386] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.627386] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.627386] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.627386] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.627956] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.627956] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.627956] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. [ 695.627956] env[61911]: ERROR nova.compute.manager [ 695.627956] env[61911]: Traceback (most recent call last): [ 695.627956] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.627956] env[61911]: listener.cb(fileno) [ 695.627956] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.627956] env[61911]: result = function(*args, **kwargs) [ 695.627956] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.627956] env[61911]: return func(*args, **kwargs) [ 695.627956] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.627956] env[61911]: raise e [ 695.627956] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.627956] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 695.627956] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.627956] env[61911]: created_port_ids = self._update_ports_for_instance( [ 695.627956] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.627956] env[61911]: with excutils.save_and_reraise_exception(): [ 695.627956] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.627956] env[61911]: self.force_reraise() [ 695.627956] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.627956] env[61911]: raise self.value [ 695.627956] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.627956] env[61911]: updated_port = self._update_port( [ 695.627956] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.627956] env[61911]: _ensure_no_port_binding_failure(port) [ 695.627956] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.627956] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.628739] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. [ 695.628739] env[61911]: Removing descriptor: 16 [ 695.628739] env[61911]: ERROR nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Traceback (most recent call last): [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] yield resources [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self.driver.spawn(context, instance, image_meta, [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.628739] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] vm_ref = self.build_virtual_machine(instance, [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] for vif in network_info: [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return self._sync_wrapper(fn, *args, **kwargs) [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self.wait() [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self[:] = self._gt.wait() [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return self._exit_event.wait() [ 695.629080] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] result = hub.switch() [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return self.greenlet.switch() [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] result = function(*args, **kwargs) [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return func(*args, **kwargs) [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] raise e [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] nwinfo = self.network_api.allocate_for_instance( [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.629420] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] created_port_ids = self._update_ports_for_instance( [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] with excutils.save_and_reraise_exception(): [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self.force_reraise() [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] raise self.value [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] updated_port = self._update_port( [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] _ensure_no_port_binding_failure(port) [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.629771] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] raise exception.PortBindingFailed(port_id=port['id']) [ 695.630091] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] nova.exception.PortBindingFailed: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. [ 695.630091] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] [ 695.630091] env[61911]: INFO nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Terminating instance [ 695.690684] env[61911]: DEBUG nova.compute.manager [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Received event network-changed-038e0939-f5ab-4535-93fb-4c4cdb1bed45 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 695.690917] env[61911]: DEBUG nova.compute.manager [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Refreshing instance network info cache due to event network-changed-038e0939-f5ab-4535-93fb-4c4cdb1bed45. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 695.691159] env[61911]: DEBUG oslo_concurrency.lockutils [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] Acquiring lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.691394] env[61911]: DEBUG oslo_concurrency.lockutils [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] Acquired lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.691456] env[61911]: DEBUG nova.network.neutron [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Refreshing network info cache for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 695.995020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Releasing lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.995020] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 695.995020] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a26171d8-f1d0-4471-b8bc-b854a6d6841b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.006745] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e758a8ba-2b37-4e38-84d8-fdd7cce65f54 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.020522] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance ade5af6e-0235-407c-85e6-9668f50ccad3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.033667] env[61911]: WARNING nova.virt.vmwareapi.driver [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance c24159b8-8974-499e-8237-26f16e3a198b could not be found. [ 696.034236] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.034236] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ef1ed32-aca5-4eef-abd1-4d9f667a882d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.042784] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1656570-16a0-4f09-9722-00c46ca1a443 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.065630] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c24159b8-8974-499e-8237-26f16e3a198b could not be found. [ 696.065630] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.065630] env[61911]: INFO nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Took 0.07 seconds to destroy the instance on the hypervisor. [ 696.065630] env[61911]: DEBUG oslo.service.loopingcall [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.065630] env[61911]: DEBUG nova.compute.manager [-] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 696.065630] env[61911]: DEBUG nova.network.neutron [-] [instance: c24159b8-8974-499e-8237-26f16e3a198b] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.086431] env[61911]: DEBUG nova.network.neutron [-] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.135814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.219687] env[61911]: DEBUG nova.network.neutron [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.293653] env[61911]: DEBUG nova.network.neutron [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.526903] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 7f777a69-d8ce-4a2f-83f5-085fad6f9809 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.589305] env[61911]: DEBUG nova.network.neutron [-] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.771546] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "28da8a83-a004-4baf-a744-bcc260032afd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.771768] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "28da8a83-a004-4baf-a744-bcc260032afd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.795764] env[61911]: DEBUG oslo_concurrency.lockutils [req-6f03ab9e-5407-4e9e-b0cf-2deccca782d5 req-a2d5b1e7-6808-4771-987d-af7a751233ba service nova] Releasing lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.796152] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquired lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.796544] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.032154] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance ea27990b-d194-4b81-b18b-37804c22ceb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.092080] env[61911]: INFO nova.compute.manager [-] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Took 1.03 seconds to deallocate network for instance. [ 697.314217] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.371917] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.535450] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.652913] env[61911]: INFO nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Took 0.56 seconds to detach 1 volumes for instance. [ 697.657061] env[61911]: DEBUG nova.compute.claims [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.657249] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.713366] env[61911]: DEBUG nova.compute.manager [req-904f52bf-1421-4e00-afcf-c4e4ba8eff27 req-3d2af3ae-e0da-4e12-bed5-93e46a976c48 service nova] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Received event network-vif-deleted-038e0939-f5ab-4535-93fb-4c4cdb1bed45 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 697.875144] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Releasing lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.875535] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 697.875726] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 697.876040] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-efddf2af-0baa-4ba2-868f-6ada57b91962 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.884922] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b5a805-f477-4edc-a246-defd8cb06fa6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.907839] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe6c9377-f9e6-43d2-afa7-2323bf60a48d could not be found. [ 697.908079] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 697.908266] env[61911]: INFO nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 697.908512] env[61911]: DEBUG oslo.service.loopingcall [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.908727] env[61911]: DEBUG nova.compute.manager [-] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 697.908837] env[61911]: DEBUG nova.network.neutron [-] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 697.924357] env[61911]: DEBUG nova.network.neutron [-] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.037904] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance fe46e0e7-c1aa-4bf5-af01-cdb04031f945 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.426876] env[61911]: DEBUG nova.network.neutron [-] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.540901] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b6773f2b-df86-4a4c-9b20-a314e81a7aa0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.929142] env[61911]: INFO nova.compute.manager [-] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Took 1.02 seconds to deallocate network for instance. [ 698.931554] env[61911]: DEBUG nova.compute.claims [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 698.931750] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.043932] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b712ffa7-19d8-47d8-8d60-48ca52735776 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.547434] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance aa8d4787-0638-4bc9-99aa-01678272f59a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.050468] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance a77a9374-0220-43ec-bf51-34523bc8babd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.554298] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance d4f17919-9cc9-4e19-88eb-38a978115975 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.058081] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 21723e9a-5c6d-4142-baba-25c59a5038ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.564552] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 66472f43-537d-4eb3-8d49-d40627a8809d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.067231] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 267075d7-c3d8-4520-b7d9-67504097f823 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.570862] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 28159e75-9fe9-44c7-b5c9-534902cecbef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.074194] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8a363b99-517c-45ff-9e33-e9ceff3a08b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.581406] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8051e8ec-6d82-416c-858a-f0cdd00f869d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.088477] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 35cf6a24-3cc0-40bd-92da-798a56b4e2c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.592062] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 951e57ba-da09-426a-b500-e5459d346f64 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.592062] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 704.592062] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 704.916141] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c83826-4a9d-49bf-84f7-1a48e29d7c2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.923656] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f83cd6-e55f-4b51-a471-6e13b40b4dc6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.953402] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115e373e-545b-45ea-b226-bf6920e9db51 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.960865] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1335e9-edea-4bfb-8be7-7ccd93ae121a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.974347] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.477257] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 705.984214] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 705.984575] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.031s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.984864] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.117s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.805923] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61053152-0c8d-463f-a5ea-1a013c21e356 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.814027] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b392ffb0-64af-4ec3-b82a-ce2665443d40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.842051] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd833f2c-0796-4ad4-91ef-0713bbdb6fe7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.849938] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9491c1a9-ba61-40ea-9d60-9a40d603c470 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.863248] env[61911]: DEBUG nova.compute.provider_tree [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.366928] env[61911]: DEBUG nova.scheduler.client.report [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 707.870948] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.886s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.871710] env[61911]: ERROR nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Traceback (most recent call last): [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self.driver.spawn(context, instance, image_meta, [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] vm_ref = self.build_virtual_machine(instance, [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.871710] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] for vif in network_info: [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return self._sync_wrapper(fn, *args, **kwargs) [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self.wait() [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self[:] = self._gt.wait() [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return self._exit_event.wait() [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] result = hub.switch() [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.872066] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return self.greenlet.switch() [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] result = function(*args, **kwargs) [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] return func(*args, **kwargs) [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] raise e [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] nwinfo = self.network_api.allocate_for_instance( [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] created_port_ids = self._update_ports_for_instance( [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] with excutils.save_and_reraise_exception(): [ 707.872419] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] self.force_reraise() [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] raise self.value [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] updated_port = self._update_port( [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] _ensure_no_port_binding_failure(port) [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] raise exception.PortBindingFailed(port_id=port['id']) [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] nova.exception.PortBindingFailed: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. [ 707.872739] env[61911]: ERROR nova.compute.manager [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] [ 707.873033] env[61911]: DEBUG nova.compute.utils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.873681] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.311s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.877513] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Build of instance f21b2aae-9d93-4d48-ba19-9905149eb5ae was re-scheduled: Binding failed for port 85e84137-d4aa-48dc-848c-4f724e1f70f7, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 707.877513] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 707.877513] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquiring lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.877513] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Acquired lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.877745] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 708.397689] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.450097] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.686025] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d422e28-b7ea-4d4d-82e5-086451e67796 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.693182] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de36dba7-f099-4b29-b357-8f9bc1069714 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.724146] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5b1650-7f1a-4038-a7e6-465d260ba607 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.731966] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e029e694-9546-4d92-a4fc-505220acada2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.747340] env[61911]: DEBUG nova.compute.provider_tree [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.952926] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Releasing lock "refresh_cache-f21b2aae-9d93-4d48-ba19-9905149eb5ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.953242] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 708.953338] env[61911]: DEBUG nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 708.953489] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.968381] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.250490] env[61911]: DEBUG nova.scheduler.client.report [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 709.473617] env[61911]: DEBUG nova.network.neutron [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.755468] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.882s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.756117] env[61911]: ERROR nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Traceback (most recent call last): [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self.driver.spawn(context, instance, image_meta, [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] vm_ref = self.build_virtual_machine(instance, [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.756117] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] for vif in network_info: [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return self._sync_wrapper(fn, *args, **kwargs) [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self.wait() [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self[:] = self._gt.wait() [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return self._exit_event.wait() [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] result = hub.switch() [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 709.756598] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return self.greenlet.switch() [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] result = function(*args, **kwargs) [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] return func(*args, **kwargs) [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] raise e [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] nwinfo = self.network_api.allocate_for_instance( [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] created_port_ids = self._update_ports_for_instance( [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] with excutils.save_and_reraise_exception(): [ 709.756985] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] self.force_reraise() [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] raise self.value [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] updated_port = self._update_port( [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] _ensure_no_port_binding_failure(port) [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] raise exception.PortBindingFailed(port_id=port['id']) [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] nova.exception.PortBindingFailed: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. [ 709.757332] env[61911]: ERROR nova.compute.manager [instance: 5235a412-c833-4f7c-a884-2855e61b8966] [ 709.757668] env[61911]: DEBUG nova.compute.utils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.758038] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.522s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.759547] env[61911]: INFO nova.compute.claims [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.762171] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Build of instance 5235a412-c833-4f7c-a884-2855e61b8966 was re-scheduled: Binding failed for port ae8c79bd-805b-46d4-bab2-81d47e27db1f, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 709.762620] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 709.762874] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Acquiring lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.762981] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Acquired lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.763229] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.979511] env[61911]: INFO nova.compute.manager [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] [instance: f21b2aae-9d93-4d48-ba19-9905149eb5ae] Took 1.02 seconds to deallocate network for instance. [ 710.281433] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.354430] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.857939] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Releasing lock "refresh_cache-5235a412-c833-4f7c-a884-2855e61b8966" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.858194] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 710.858379] env[61911]: DEBUG nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 710.858545] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.873455] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.016624] env[61911]: INFO nova.scheduler.client.report [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Deleted allocations for instance f21b2aae-9d93-4d48-ba19-9905149eb5ae [ 711.096936] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c761d7c4-48ae-4a9b-bd35-9958aa9c21cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.104787] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04bab7d-5cc0-40bd-ae5d-83d571f173e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.134550] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b104c31-25aa-40d7-94f8-fa569c820e4b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.141330] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d2d0e0-3d31-4285-92a9-948a20d7950f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.153978] env[61911]: DEBUG nova.compute.provider_tree [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.376209] env[61911]: DEBUG nova.network.neutron [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.525862] env[61911]: DEBUG oslo_concurrency.lockutils [None req-da1cdded-88f7-49d2-9e0a-9cc34342db82 tempest-DeleteServersAdminTestJSON-1770441033 tempest-DeleteServersAdminTestJSON-1770441033-project-member] Lock "f21b2aae-9d93-4d48-ba19-9905149eb5ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.881s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.657707] env[61911]: DEBUG nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 711.881946] env[61911]: INFO nova.compute.manager [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] [instance: 5235a412-c833-4f7c-a884-2855e61b8966] Took 1.02 seconds to deallocate network for instance. [ 712.028613] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 712.162793] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.163371] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 712.165800] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.595s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.551214] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.670284] env[61911]: DEBUG nova.compute.utils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.674725] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 712.674910] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.728275] env[61911]: DEBUG nova.policy [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a75ff7724318435b9b4a733df8edde2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d55061be50947d3a1d72d29047c2312', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.915668] env[61911]: INFO nova.scheduler.client.report [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Deleted allocations for instance 5235a412-c833-4f7c-a884-2855e61b8966 [ 713.026944] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Successfully created port: 23386ee7-1825-4aa5-9065-66ddfcf27500 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.087371] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3401d8d-dcb3-42e2-a92b-451950fe89da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.096021] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc80ac2-6d0c-4f6c-9589-fe582361d71a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.139124] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b223ca0e-3a79-46c8-9b7a-022a31bba040 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.147294] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a7b09b-c567-425b-baca-bd96877ac038 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.162647] env[61911]: DEBUG nova.compute.provider_tree [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.175689] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 713.436645] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c929201f-29ed-4d16-8e2d-bc9295ad0c8e tempest-ImagesOneServerTestJSON-260745187 tempest-ImagesOneServerTestJSON-260745187-project-member] Lock "5235a412-c833-4f7c-a884-2855e61b8966" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.518s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.665484] env[61911]: DEBUG nova.scheduler.client.report [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 713.704702] env[61911]: DEBUG nova.compute.manager [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Received event network-changed-23386ee7-1825-4aa5-9065-66ddfcf27500 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 713.704906] env[61911]: DEBUG nova.compute.manager [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Refreshing instance network info cache due to event network-changed-23386ee7-1825-4aa5-9065-66ddfcf27500. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 713.705142] env[61911]: DEBUG oslo_concurrency.lockutils [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] Acquiring lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.705289] env[61911]: DEBUG oslo_concurrency.lockutils [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] Acquired lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.705454] env[61911]: DEBUG nova.network.neutron [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Refreshing network info cache for port 23386ee7-1825-4aa5-9065-66ddfcf27500 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.878629] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. [ 713.878629] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.878629] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.878629] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.878629] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.878629] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.878629] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.878629] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.878629] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.878629] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 713.878629] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.878629] env[61911]: ERROR nova.compute.manager raise self.value [ 713.878629] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.878629] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.878629] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.878629] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.879109] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.879109] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.879109] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. [ 713.879109] env[61911]: ERROR nova.compute.manager [ 713.879109] env[61911]: Traceback (most recent call last): [ 713.879109] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.879109] env[61911]: listener.cb(fileno) [ 713.879109] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.879109] env[61911]: result = function(*args, **kwargs) [ 713.879109] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.879109] env[61911]: return func(*args, **kwargs) [ 713.879109] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 713.879109] env[61911]: raise e [ 713.879109] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.879109] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 713.879109] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.879109] env[61911]: created_port_ids = self._update_ports_for_instance( [ 713.879109] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.879109] env[61911]: with excutils.save_and_reraise_exception(): [ 713.879109] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.879109] env[61911]: self.force_reraise() [ 713.879109] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.879109] env[61911]: raise self.value [ 713.879109] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.879109] env[61911]: updated_port = self._update_port( [ 713.879109] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.879109] env[61911]: _ensure_no_port_binding_failure(port) [ 713.879109] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.879109] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.879836] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. [ 713.879836] env[61911]: Removing descriptor: 16 [ 713.939622] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 714.172630] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.173297] env[61911]: ERROR nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Traceback (most recent call last): [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self.driver.spawn(context, instance, image_meta, [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] vm_ref = self.build_virtual_machine(instance, [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.173297] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] for vif in network_info: [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] return self._sync_wrapper(fn, *args, **kwargs) [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self.wait() [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self[:] = self._gt.wait() [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] return self._exit_event.wait() [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] current.throw(*self._exc) [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.173872] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] result = function(*args, **kwargs) [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] return func(*args, **kwargs) [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] raise e [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] nwinfo = self.network_api.allocate_for_instance( [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] created_port_ids = self._update_ports_for_instance( [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] with excutils.save_and_reraise_exception(): [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] self.force_reraise() [ 714.174415] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] raise self.value [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] updated_port = self._update_port( [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] _ensure_no_port_binding_failure(port) [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] raise exception.PortBindingFailed(port_id=port['id']) [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] nova.exception.PortBindingFailed: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. [ 714.174960] env[61911]: ERROR nova.compute.manager [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] [ 714.174960] env[61911]: DEBUG nova.compute.utils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.175554] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.368s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.177945] env[61911]: INFO nova.compute.claims [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.182143] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Build of instance 5002bdba-3519-4caa-8486-c2249bd9abef was re-scheduled: Binding failed for port edbb6d57-578f-4232-ad87-d8968e20e05e, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 714.182594] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 714.182822] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Acquiring lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.182969] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Acquired lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.183461] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.187946] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 714.218427] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 714.218680] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.218838] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 714.219028] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.219176] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 714.219358] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 714.219631] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 714.219771] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 714.219938] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 714.220148] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 714.220332] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 714.221713] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bf38e4-0515-4f0a-a93c-d21e9480a899 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.231965] env[61911]: DEBUG nova.network.neutron [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.234754] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27b84f9-5e22-48a5-9c07-90207e473592 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.252427] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Traceback (most recent call last): [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] yield resources [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self.driver.spawn(context, instance, image_meta, [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] vm_ref = self.build_virtual_machine(instance, [ 714.252427] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] for vif in network_info: [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] return self._sync_wrapper(fn, *args, **kwargs) [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self.wait() [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self[:] = self._gt.wait() [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] return self._exit_event.wait() [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.253030] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] current.throw(*self._exc) [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] result = function(*args, **kwargs) [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] return func(*args, **kwargs) [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] raise e [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] nwinfo = self.network_api.allocate_for_instance( [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] created_port_ids = self._update_ports_for_instance( [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] with excutils.save_and_reraise_exception(): [ 714.253588] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self.force_reraise() [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] raise self.value [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] updated_port = self._update_port( [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] _ensure_no_port_binding_failure(port) [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] raise exception.PortBindingFailed(port_id=port['id']) [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] nova.exception.PortBindingFailed: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. [ 714.253916] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] [ 714.253916] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Terminating instance [ 714.375168] env[61911]: DEBUG nova.network.neutron [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.465791] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.707945] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.758217] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.820073] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.878095] env[61911]: DEBUG oslo_concurrency.lockutils [req-4aa813fd-70ba-41ab-a80a-aa5edb84e10c req-be8c5b05-f350-4cbd-89ba-d581c4ac934f service nova] Releasing lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.878095] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquired lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.879134] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.321344] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Releasing lock "refresh_cache-5002bdba-3519-4caa-8486-c2249bd9abef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.321609] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 715.324020] env[61911]: DEBUG nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 715.324020] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.343832] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.559206] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.638132] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.655222] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d3a289-e6c8-40cf-84f9-4239f24c20e2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.666443] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190cdc17-4736-4337-8160-3a21aa7d8932 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.706394] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1edfb3cc-c048-481a-9382-53772152cce7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.714402] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41fd28b-7bda-4864-ad37-2e1147b7c039 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.729835] env[61911]: DEBUG nova.compute.provider_tree [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.846528] env[61911]: DEBUG nova.network.neutron [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.875428] env[61911]: DEBUG nova.compute.manager [req-151a8397-0181-431d-931c-8ecd63884f8a req-031df808-3226-4939-8610-e428c42a9b96 service nova] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Received event network-vif-deleted-23386ee7-1825-4aa5-9065-66ddfcf27500 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 716.144547] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "43d18895-202c-4048-9435-b3484ffd4c07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.145925] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.146331] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Releasing lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.146690] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 716.146873] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 716.147401] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f474fd3-efba-4b3e-a9f0-525ad70441be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.156534] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d598e140-4d24-4017-bf67-d306e99d394d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.180505] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e5f4224f-0002-4b75-9f69-d1ac50d955ae could not be found. [ 716.180744] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.180924] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Took 0.03 seconds to destroy the instance on the hypervisor. [ 716.181185] env[61911]: DEBUG oslo.service.loopingcall [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.181798] env[61911]: DEBUG nova.compute.manager [-] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 716.181798] env[61911]: DEBUG nova.network.neutron [-] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.202639] env[61911]: DEBUG nova.network.neutron [-] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.234831] env[61911]: DEBUG nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 716.350351] env[61911]: INFO nova.compute.manager [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] [instance: 5002bdba-3519-4caa-8486-c2249bd9abef] Took 1.03 seconds to deallocate network for instance. [ 716.705092] env[61911]: DEBUG nova.network.neutron [-] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.739388] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.739925] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 716.742872] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.156s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.208014] env[61911]: INFO nova.compute.manager [-] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Took 1.03 seconds to deallocate network for instance. [ 717.210953] env[61911]: DEBUG nova.compute.claims [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 717.211151] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.245824] env[61911]: DEBUG nova.compute.utils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.246157] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 717.246327] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 717.294502] env[61911]: DEBUG nova.policy [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a75ff7724318435b9b4a733df8edde2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d55061be50947d3a1d72d29047c2312', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.386840] env[61911]: INFO nova.scheduler.client.report [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Deleted allocations for instance 5002bdba-3519-4caa-8486-c2249bd9abef [ 717.608891] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Successfully created port: 7c19d4a7-07f8-4102-bc77-8893acbb3672 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.647910] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebff43e-6b24-4736-805e-e33dec9b083d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.655940] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772b0984-a961-48c5-a119-5742033fa294 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.689950] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5742192-798b-47f1-8a19-cb98fd494d3d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.698919] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bb4033-e3a5-4e84-85d2-ec4f9b2182a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.712996] env[61911]: DEBUG nova.compute.provider_tree [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.749600] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 717.902117] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4177c7df-5a24-440a-8d56-5f4a08baba67 tempest-InstanceActionsV221TestJSON-464838657 tempest-InstanceActionsV221TestJSON-464838657-project-member] Lock "5002bdba-3519-4caa-8486-c2249bd9abef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.788s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.216246] env[61911]: DEBUG nova.scheduler.client.report [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 718.317390] env[61911]: DEBUG nova.compute.manager [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Received event network-changed-7c19d4a7-07f8-4102-bc77-8893acbb3672 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 718.317390] env[61911]: DEBUG nova.compute.manager [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Refreshing instance network info cache due to event network-changed-7c19d4a7-07f8-4102-bc77-8893acbb3672. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 718.317390] env[61911]: DEBUG oslo_concurrency.lockutils [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] Acquiring lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.317390] env[61911]: DEBUG oslo_concurrency.lockutils [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] Acquired lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.317390] env[61911]: DEBUG nova.network.neutron [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Refreshing network info cache for port 7c19d4a7-07f8-4102-bc77-8893acbb3672 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.404989] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 718.505119] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. [ 718.505119] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.505119] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.505119] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.505119] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.505119] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.505119] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.505119] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.505119] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.505119] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 718.505119] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.505119] env[61911]: ERROR nova.compute.manager raise self.value [ 718.505119] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.505119] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.505119] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.505119] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.505613] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.505613] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.505613] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. [ 718.505613] env[61911]: ERROR nova.compute.manager [ 718.505613] env[61911]: Traceback (most recent call last): [ 718.505613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.505613] env[61911]: listener.cb(fileno) [ 718.505613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.505613] env[61911]: result = function(*args, **kwargs) [ 718.505613] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.505613] env[61911]: return func(*args, **kwargs) [ 718.505613] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.505613] env[61911]: raise e [ 718.505613] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.505613] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 718.505613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.505613] env[61911]: created_port_ids = self._update_ports_for_instance( [ 718.505613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.505613] env[61911]: with excutils.save_and_reraise_exception(): [ 718.505613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.505613] env[61911]: self.force_reraise() [ 718.505613] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.505613] env[61911]: raise self.value [ 718.505613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.505613] env[61911]: updated_port = self._update_port( [ 718.505613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.505613] env[61911]: _ensure_no_port_binding_failure(port) [ 718.505613] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.505613] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.506377] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. [ 718.506377] env[61911]: Removing descriptor: 16 [ 718.720895] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.721694] env[61911]: ERROR nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Traceback (most recent call last): [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self.driver.spawn(context, instance, image_meta, [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] vm_ref = self.build_virtual_machine(instance, [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.721694] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] for vif in network_info: [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] return self._sync_wrapper(fn, *args, **kwargs) [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self.wait() [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self[:] = self._gt.wait() [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] return self._exit_event.wait() [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] current.throw(*self._exc) [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.722132] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] result = function(*args, **kwargs) [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] return func(*args, **kwargs) [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] raise e [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] nwinfo = self.network_api.allocate_for_instance( [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] created_port_ids = self._update_ports_for_instance( [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] with excutils.save_and_reraise_exception(): [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] self.force_reraise() [ 718.722502] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] raise self.value [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] updated_port = self._update_port( [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] _ensure_no_port_binding_failure(port) [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] raise exception.PortBindingFailed(port_id=port['id']) [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] nova.exception.PortBindingFailed: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. [ 718.722879] env[61911]: ERROR nova.compute.manager [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] [ 718.722879] env[61911]: DEBUG nova.compute.utils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.723690] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.550s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.725109] env[61911]: INFO nova.compute.claims [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.727537] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Build of instance 61c4a92c-296a-4c0e-aa9a-32b20a837b6d was re-scheduled: Binding failed for port 8bbf43c0-a8da-4c8d-ab23-d3e6afdc6e0a, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 718.727952] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 718.728187] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.728335] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.728494] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.762806] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 718.789421] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 718.789670] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.789826] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.790013] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.790175] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.790321] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 718.790527] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 718.790686] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 718.790849] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 718.791014] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 718.791190] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 718.792102] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7420f39-6817-4bad-8fa6-cd0824f039e2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.800278] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0898487e-b801-4010-a6a3-499aef9eb97d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.814039] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Traceback (most recent call last): [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] yield resources [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self.driver.spawn(context, instance, image_meta, [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] vm_ref = self.build_virtual_machine(instance, [ 718.814039] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] for vif in network_info: [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] return self._sync_wrapper(fn, *args, **kwargs) [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self.wait() [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self[:] = self._gt.wait() [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] return self._exit_event.wait() [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.814481] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] current.throw(*self._exc) [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] result = function(*args, **kwargs) [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] return func(*args, **kwargs) [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] raise e [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] nwinfo = self.network_api.allocate_for_instance( [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] created_port_ids = self._update_ports_for_instance( [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] with excutils.save_and_reraise_exception(): [ 718.814863] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self.force_reraise() [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] raise self.value [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] updated_port = self._update_port( [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] _ensure_no_port_binding_failure(port) [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] raise exception.PortBindingFailed(port_id=port['id']) [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] nova.exception.PortBindingFailed: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. [ 718.815240] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] [ 718.815240] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Terminating instance [ 718.836290] env[61911]: DEBUG nova.network.neutron [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.916857] env[61911]: DEBUG nova.network.neutron [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.936092] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.256177] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.317927] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.335497] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.421686] env[61911]: DEBUG oslo_concurrency.lockutils [req-038df9a1-888e-4c55-ab3d-2f742e750b36 req-8af5a594-4475-4252-b1b9-292c6bba142d service nova] Releasing lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.422103] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquired lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.422293] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.840775] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-61c4a92c-296a-4c0e-aa9a-32b20a837b6d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.841016] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 719.841203] env[61911]: DEBUG nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 719.841365] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.862213] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.938093] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.024284] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.135995] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ff3dde-a977-4275-b685-c2025ce0856a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.143811] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f43f8b-4d3b-44b5-af37-21425e0f6e96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.174160] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dd811b-3bc1-4e4e-acba-89c6769f89dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.182122] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a380941b-1668-450e-95a3-dc408947372b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.197910] env[61911]: DEBUG nova.compute.provider_tree [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.365599] env[61911]: DEBUG nova.network.neutron [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.426330] env[61911]: DEBUG nova.compute.manager [req-0534d85e-d7a0-4bea-8694-b972c7d7e416 req-c0d37aad-6216-4e0d-ad0d-7f530ea8da2b service nova] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Received event network-vif-deleted-7c19d4a7-07f8-4102-bc77-8893acbb3672 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 720.527358] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Releasing lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.527801] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 720.528100] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.528398] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f329e05-ce4f-43ab-9a47-83d91aadccec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.537177] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6ec2c7-416c-4a3c-a3c5-7dccad9236b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.559021] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325 could not be found. [ 720.559249] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.559425] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Took 0.03 seconds to destroy the instance on the hypervisor. [ 720.559666] env[61911]: DEBUG oslo.service.loopingcall [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.559876] env[61911]: DEBUG nova.compute.manager [-] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 720.559961] env[61911]: DEBUG nova.network.neutron [-] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.577385] env[61911]: DEBUG nova.network.neutron [-] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.700703] env[61911]: DEBUG nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 720.869800] env[61911]: INFO nova.compute.manager [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: 61c4a92c-296a-4c0e-aa9a-32b20a837b6d] Took 1.03 seconds to deallocate network for instance. [ 721.079689] env[61911]: DEBUG nova.network.neutron [-] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.206820] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.207965] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 721.211046] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.121s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.582585] env[61911]: INFO nova.compute.manager [-] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Took 1.02 seconds to deallocate network for instance. [ 721.585078] env[61911]: DEBUG nova.compute.claims [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.585284] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.720461] env[61911]: DEBUG nova.compute.utils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.724917] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 721.725027] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.797019] env[61911]: DEBUG nova.policy [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a75ff7724318435b9b4a733df8edde2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d55061be50947d3a1d72d29047c2312', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.896986] env[61911]: INFO nova.scheduler.client.report [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleted allocations for instance 61c4a92c-296a-4c0e-aa9a-32b20a837b6d [ 722.065854] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9698243-e043-4ad4-88de-4d3d6d51d027 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.072574] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e427dc-84f3-42e7-a89a-25bfee0e59c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.102031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c403a03a-7e2e-4577-888c-79203431daff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.109657] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0795eb7d-f5f3-47c0-9e65-589708229f76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.125446] env[61911]: DEBUG nova.compute.provider_tree [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.137572] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Successfully created port: 8486f934-8057-4b47-9f6b-4449b35edb4e {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.228435] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 722.406771] env[61911]: DEBUG oslo_concurrency.lockutils [None req-50cff4a8-1711-42da-a1ff-6b3c10d6c5a9 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "61c4a92c-296a-4c0e-aa9a-32b20a837b6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.135s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.631561] env[61911]: DEBUG nova.scheduler.client.report [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 722.911507] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 722.969737] env[61911]: DEBUG nova.compute.manager [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Received event network-changed-8486f934-8057-4b47-9f6b-4449b35edb4e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 722.969737] env[61911]: DEBUG nova.compute.manager [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Refreshing instance network info cache due to event network-changed-8486f934-8057-4b47-9f6b-4449b35edb4e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 722.969946] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] Acquiring lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.970090] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] Acquired lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.971937] env[61911]: DEBUG nova.network.neutron [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Refreshing network info cache for port 8486f934-8057-4b47-9f6b-4449b35edb4e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 723.027418] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. [ 723.027418] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.027418] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 723.027418] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.027418] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.027418] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.027418] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.027418] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.027418] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.027418] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 723.027418] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.027418] env[61911]: ERROR nova.compute.manager raise self.value [ 723.027418] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.027418] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.027418] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.027418] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.027828] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.027828] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.027828] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. [ 723.027828] env[61911]: ERROR nova.compute.manager [ 723.027828] env[61911]: Traceback (most recent call last): [ 723.027828] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.027828] env[61911]: listener.cb(fileno) [ 723.027828] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.027828] env[61911]: result = function(*args, **kwargs) [ 723.027828] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.027828] env[61911]: return func(*args, **kwargs) [ 723.027828] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 723.027828] env[61911]: raise e [ 723.027828] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 723.027828] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 723.027828] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.027828] env[61911]: created_port_ids = self._update_ports_for_instance( [ 723.027828] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.027828] env[61911]: with excutils.save_and_reraise_exception(): [ 723.027828] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.027828] env[61911]: self.force_reraise() [ 723.027828] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.027828] env[61911]: raise self.value [ 723.027828] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.027828] env[61911]: updated_port = self._update_port( [ 723.027828] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.027828] env[61911]: _ensure_no_port_binding_failure(port) [ 723.027828] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.027828] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.028547] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. [ 723.028547] env[61911]: Removing descriptor: 16 [ 723.136616] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.925s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.137269] env[61911]: ERROR nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Traceback (most recent call last): [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self.driver.spawn(context, instance, image_meta, [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] vm_ref = self.build_virtual_machine(instance, [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.137269] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] for vif in network_info: [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] return self._sync_wrapper(fn, *args, **kwargs) [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self.wait() [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self[:] = self._gt.wait() [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] return self._exit_event.wait() [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] current.throw(*self._exc) [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.137582] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] result = function(*args, **kwargs) [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] return func(*args, **kwargs) [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] raise e [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] nwinfo = self.network_api.allocate_for_instance( [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] created_port_ids = self._update_ports_for_instance( [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] with excutils.save_and_reraise_exception(): [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] self.force_reraise() [ 723.137904] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] raise self.value [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] updated_port = self._update_port( [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] _ensure_no_port_binding_failure(port) [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] raise exception.PortBindingFailed(port_id=port['id']) [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] nova.exception.PortBindingFailed: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. [ 723.138237] env[61911]: ERROR nova.compute.manager [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] [ 723.138237] env[61911]: DEBUG nova.compute.utils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.139751] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.482s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.145659] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Build of instance 08e38cae-cf76-4367-ac66-c5ecd25c2763 was re-scheduled: Binding failed for port c6e44c9b-bfe8-4463-8444-6ebae2c90be6, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 723.145659] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 723.145659] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Acquiring lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.145659] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Acquired lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.146025] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.240770] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 723.281075] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 723.281362] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.281548] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 723.282326] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.282326] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 723.282574] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 723.282922] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 723.282922] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 723.283158] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 723.283397] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 723.286092] env[61911]: DEBUG nova.virt.hardware [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 723.286092] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c60bfb9-b133-41af-b49b-203b7fcdf8bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.294992] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf7d7ad-70ee-47a3-af3d-0462663404d1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.310173] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Traceback (most recent call last): [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] yield resources [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self.driver.spawn(context, instance, image_meta, [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] vm_ref = self.build_virtual_machine(instance, [ 723.310173] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] for vif in network_info: [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] return self._sync_wrapper(fn, *args, **kwargs) [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self.wait() [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self[:] = self._gt.wait() [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] return self._exit_event.wait() [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 723.310535] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] current.throw(*self._exc) [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] result = function(*args, **kwargs) [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] return func(*args, **kwargs) [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] raise e [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] nwinfo = self.network_api.allocate_for_instance( [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] created_port_ids = self._update_ports_for_instance( [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] with excutils.save_and_reraise_exception(): [ 723.310891] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self.force_reraise() [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] raise self.value [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] updated_port = self._update_port( [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] _ensure_no_port_binding_failure(port) [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] raise exception.PortBindingFailed(port_id=port['id']) [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] nova.exception.PortBindingFailed: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. [ 723.311258] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] [ 723.311258] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Terminating instance [ 723.438201] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.494474] env[61911]: DEBUG nova.network.neutron [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.589661] env[61911]: DEBUG nova.network.neutron [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.668330] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.748037] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.815295] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.067214] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b083d8cc-e14c-4754-b7e7-911cd2c574b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.077658] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0ebda8-c9ed-41f7-b59c-4bb4a8095911 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.113685] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6ae36a2-7a8b-4554-8f32-ee3b848fc3ca req-143fe27b-65bc-4e2e-9145-b939466727aa service nova] Releasing lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.113822] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquired lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.114048] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.116148] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfc544e-13c7-4ec9-a400-15008badb26c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.124642] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d7ee2c-6056-44f3-89f8-10edf35b93ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.141523] env[61911]: DEBUG nova.compute.provider_tree [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.251108] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Releasing lock "refresh_cache-08e38cae-cf76-4367-ac66-c5ecd25c2763" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.254544] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 724.254544] env[61911]: DEBUG nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 724.254544] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.270680] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.496768] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "31f0a4f7-a69d-4ec4-966d-93f402163b88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.497015] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.635334] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.644220] env[61911]: DEBUG nova.scheduler.client.report [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 724.712989] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.772732] env[61911]: DEBUG nova.network.neutron [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.028626] env[61911]: DEBUG nova.compute.manager [req-76396c8a-4ebf-469e-8f72-ef098509a58a req-1d991805-a2ab-4d6f-931d-59f612ea1e8f service nova] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Received event network-vif-deleted-8486f934-8057-4b47-9f6b-4449b35edb4e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 725.149150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.149689] env[61911]: ERROR nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] Traceback (most recent call last): [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self.driver.spawn(context, instance, image_meta, [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] vm_ref = self.build_virtual_machine(instance, [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.149689] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] for vif in network_info: [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] return self._sync_wrapper(fn, *args, **kwargs) [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self.wait() [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self[:] = self._gt.wait() [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] return self._exit_event.wait() [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] current.throw(*self._exc) [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.150030] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] result = function(*args, **kwargs) [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] return func(*args, **kwargs) [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] raise e [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] nwinfo = self.network_api.allocate_for_instance( [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] created_port_ids = self._update_ports_for_instance( [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] with excutils.save_and_reraise_exception(): [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] self.force_reraise() [ 725.150346] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] raise self.value [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] updated_port = self._update_port( [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] _ensure_no_port_binding_failure(port) [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] raise exception.PortBindingFailed(port_id=port['id']) [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] nova.exception.PortBindingFailed: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. [ 725.150752] env[61911]: ERROR nova.compute.manager [instance: c24159b8-8974-499e-8237-26f16e3a198b] [ 725.150752] env[61911]: DEBUG nova.compute.utils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.151852] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.220s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.155208] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Build of instance c24159b8-8974-499e-8237-26f16e3a198b was re-scheduled: Binding failed for port b4c68e4c-f488-40d7-be84-76f26f71daa3, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 725.155350] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 725.156038] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Acquiring lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.156038] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Acquired lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.156038] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.215651] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Releasing lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.216099] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 725.216298] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.216604] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40caec08-9055-4a88-94c0-0aa3a6a7e8b6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.228046] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9efee44-7191-45ce-887f-610765a74be1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.251277] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a09bc118-c3f8-4943-89a1-8a0c59d528c5 could not be found. [ 725.251505] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.251799] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 725.252088] env[61911]: DEBUG oslo.service.loopingcall [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.252351] env[61911]: DEBUG nova.compute.manager [-] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 725.252472] env[61911]: DEBUG nova.network.neutron [-] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.267834] env[61911]: DEBUG nova.network.neutron [-] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.275269] env[61911]: INFO nova.compute.manager [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] [instance: 08e38cae-cf76-4367-ac66-c5ecd25c2763] Took 1.02 seconds to deallocate network for instance. [ 725.577445] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.577876] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.676526] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.762611] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.770973] env[61911]: DEBUG nova.network.neutron [-] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.980927] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cc1a0a-f477-4db5-8f3c-4a8bc4daf9bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.988548] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac545290-4b8d-4a3b-b0a9-6f4ef6d505ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.018292] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14378ba9-c360-4d00-b724-7a257fa7a207 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.026034] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8c049d-716d-47f4-9c9f-529b757523f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.038794] env[61911]: DEBUG nova.compute.provider_tree [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.267695] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Releasing lock "refresh_cache-c24159b8-8974-499e-8237-26f16e3a198b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.267695] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 726.267695] env[61911]: DEBUG nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 726.267695] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.273895] env[61911]: INFO nova.compute.manager [-] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Took 1.02 seconds to deallocate network for instance. [ 726.278267] env[61911]: DEBUG nova.compute.claims [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.278267] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.286149] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.314019] env[61911]: INFO nova.scheduler.client.report [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Deleted allocations for instance 08e38cae-cf76-4367-ac66-c5ecd25c2763 [ 726.543095] env[61911]: DEBUG nova.scheduler.client.report [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 726.788819] env[61911]: DEBUG nova.network.neutron [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.822747] env[61911]: DEBUG oslo_concurrency.lockutils [None req-84cd4d71-592f-4bc3-9c35-06e6b3efd3ce tempest-ServerActionsV293TestJSON-2009312576 tempest-ServerActionsV293TestJSON-2009312576-project-member] Lock "08e38cae-cf76-4367-ac66-c5ecd25c2763" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.863s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.047850] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.048554] env[61911]: ERROR nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Traceback (most recent call last): [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self.driver.spawn(context, instance, image_meta, [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] vm_ref = self.build_virtual_machine(instance, [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.048554] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] for vif in network_info: [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return self._sync_wrapper(fn, *args, **kwargs) [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self.wait() [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self[:] = self._gt.wait() [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return self._exit_event.wait() [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] result = hub.switch() [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 727.048883] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return self.greenlet.switch() [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] result = function(*args, **kwargs) [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] return func(*args, **kwargs) [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] raise e [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] nwinfo = self.network_api.allocate_for_instance( [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] created_port_ids = self._update_ports_for_instance( [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] with excutils.save_and_reraise_exception(): [ 727.049618] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] self.force_reraise() [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] raise self.value [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] updated_port = self._update_port( [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] _ensure_no_port_binding_failure(port) [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] raise exception.PortBindingFailed(port_id=port['id']) [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] nova.exception.PortBindingFailed: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. [ 727.050206] env[61911]: ERROR nova.compute.manager [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] [ 727.050684] env[61911]: DEBUG nova.compute.utils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 727.050684] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.500s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.052874] env[61911]: INFO nova.compute.claims [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.055027] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Build of instance fe6c9377-f9e6-43d2-afa7-2323bf60a48d was re-scheduled: Binding failed for port 038e0939-f5ab-4535-93fb-4c4cdb1bed45, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 727.055327] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 727.055554] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquiring lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.055700] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Acquired lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.055859] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.292775] env[61911]: INFO nova.compute.manager [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] [instance: c24159b8-8974-499e-8237-26f16e3a198b] Took 1.03 seconds to deallocate network for instance. [ 727.325727] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 727.576707] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.625943] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.847115] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.128249] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Releasing lock "refresh_cache-fe6c9377-f9e6-43d2-afa7-2323bf60a48d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.128492] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 728.128658] env[61911]: DEBUG nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 728.128829] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.145746] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.324989] env[61911]: INFO nova.scheduler.client.report [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Deleted allocations for instance c24159b8-8974-499e-8237-26f16e3a198b [ 728.465050] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c86f70-1b34-4591-a88f-f18af2f97121 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.474902] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e32447-472a-4423-8c80-d8ae37aa64ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.505757] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0229aaa-8f48-4c0c-baa3-a5ca76493aa4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.513402] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77eafafc-fb17-475b-ab95-b88cabc4f98d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.527133] env[61911]: DEBUG nova.compute.provider_tree [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.649597] env[61911]: DEBUG nova.network.neutron [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.835900] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ac6c35fa-f3d3-44c1-9b38-22ef4cb1d274 tempest-ServersTestBootFromVolume-1097289715 tempest-ServersTestBootFromVolume-1097289715-project-member] Lock "c24159b8-8974-499e-8237-26f16e3a198b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.751s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.034055] env[61911]: DEBUG nova.scheduler.client.report [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 729.152904] env[61911]: INFO nova.compute.manager [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] [instance: fe6c9377-f9e6-43d2-afa7-2323bf60a48d] Took 1.02 seconds to deallocate network for instance. [ 729.341638] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 729.540297] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.540297] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 729.541761] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.076s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.543346] env[61911]: INFO nova.compute.claims [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.863909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.049525] env[61911]: DEBUG nova.compute.utils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.056621] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 730.056621] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 730.109625] env[61911]: DEBUG nova.policy [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dde001e8a06c4008a0e43dddc4132514', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0684fe0594a84f10a687e610a7b48fe3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.197496] env[61911]: INFO nova.scheduler.client.report [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Deleted allocations for instance fe6c9377-f9e6-43d2-afa7-2323bf60a48d [ 730.462868] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Successfully created port: 2eb47a66-a7b9-4232-982b-6b27c45817f5 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.554431] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 730.706596] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c59ded9-7eea-472f-9f03-14e65d1ba69f tempest-MigrationsAdminTest-303834964 tempest-MigrationsAdminTest-303834964-project-member] Lock "fe6c9377-f9e6-43d2-afa7-2323bf60a48d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.022s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.930411] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e362fa4e-67f2-4ff3-9ae7-65aa1a5da8ed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.940162] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f676f84c-f4e5-4027-83a7-24de6b365b3a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.979867] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4be52e3-6e2b-4532-9cb0-5a1607937368 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.990102] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61437489-c71d-475c-92b8-1d0047ec0412 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.006059] env[61911]: DEBUG nova.compute.provider_tree [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.214635] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 731.274924] env[61911]: DEBUG nova.compute.manager [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Received event network-changed-2eb47a66-a7b9-4232-982b-6b27c45817f5 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 731.275393] env[61911]: DEBUG nova.compute.manager [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Refreshing instance network info cache due to event network-changed-2eb47a66-a7b9-4232-982b-6b27c45817f5. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 731.275793] env[61911]: DEBUG oslo_concurrency.lockutils [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] Acquiring lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.277138] env[61911]: DEBUG oslo_concurrency.lockutils [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] Acquired lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.277138] env[61911]: DEBUG nova.network.neutron [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Refreshing network info cache for port 2eb47a66-a7b9-4232-982b-6b27c45817f5 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 731.512207] env[61911]: DEBUG nova.scheduler.client.report [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 731.534712] env[61911]: ERROR nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. [ 731.534712] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 731.534712] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 731.534712] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 731.534712] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.534712] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 731.534712] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.534712] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 731.534712] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.534712] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 731.534712] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.534712] env[61911]: ERROR nova.compute.manager raise self.value [ 731.534712] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.534712] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 731.534712] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.534712] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 731.535169] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.535169] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 731.535169] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. [ 731.535169] env[61911]: ERROR nova.compute.manager [ 731.535169] env[61911]: Traceback (most recent call last): [ 731.535169] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 731.535169] env[61911]: listener.cb(fileno) [ 731.535169] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.535169] env[61911]: result = function(*args, **kwargs) [ 731.535169] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.535169] env[61911]: return func(*args, **kwargs) [ 731.535169] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 731.535169] env[61911]: raise e [ 731.535169] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 731.535169] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 731.535169] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.535169] env[61911]: created_port_ids = self._update_ports_for_instance( [ 731.535169] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.535169] env[61911]: with excutils.save_and_reraise_exception(): [ 731.535169] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.535169] env[61911]: self.force_reraise() [ 731.535169] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.535169] env[61911]: raise self.value [ 731.535169] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.535169] env[61911]: updated_port = self._update_port( [ 731.535169] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.535169] env[61911]: _ensure_no_port_binding_failure(port) [ 731.535169] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.535169] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 731.535997] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. [ 731.535997] env[61911]: Removing descriptor: 16 [ 731.567377] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 731.592126] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 731.592402] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.592560] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 731.592745] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.592890] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 731.593139] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 731.593415] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 731.593610] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 731.593816] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 731.594053] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 731.594277] env[61911]: DEBUG nova.virt.hardware [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 731.595208] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fec7010-00d4-46e5-aa60-79808b8352bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.604235] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359ea2de-466e-4791-89fc-4570a89d4f6f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.618882] env[61911]: ERROR nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Traceback (most recent call last): [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] yield resources [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self.driver.spawn(context, instance, image_meta, [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] vm_ref = self.build_virtual_machine(instance, [ 731.618882] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] for vif in network_info: [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] return self._sync_wrapper(fn, *args, **kwargs) [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self.wait() [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self[:] = self._gt.wait() [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] return self._exit_event.wait() [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 731.619326] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] current.throw(*self._exc) [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] result = function(*args, **kwargs) [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] return func(*args, **kwargs) [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] raise e [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] nwinfo = self.network_api.allocate_for_instance( [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] created_port_ids = self._update_ports_for_instance( [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] with excutils.save_and_reraise_exception(): [ 731.619825] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self.force_reraise() [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] raise self.value [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] updated_port = self._update_port( [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] _ensure_no_port_binding_failure(port) [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] raise exception.PortBindingFailed(port_id=port['id']) [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] nova.exception.PortBindingFailed: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. [ 731.620319] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] [ 731.620319] env[61911]: INFO nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Terminating instance [ 731.750464] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.810421] env[61911]: DEBUG nova.network.neutron [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.952794] env[61911]: DEBUG nova.network.neutron [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.019969] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.019969] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 732.020744] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.810s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.125373] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Acquiring lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.456141] env[61911]: DEBUG oslo_concurrency.lockutils [req-bb867dd2-5e6f-42bc-a10d-7f353755e9a4 req-32688e10-e317-4a0a-a889-b9ea86508e0e service nova] Releasing lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.457021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Acquired lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.457021] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.526611] env[61911]: DEBUG nova.compute.utils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.532148] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 732.532428] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 732.590701] env[61911]: DEBUG nova.policy [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc491989e36941f2bec32d605123d1bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2a4075ba2c546eca3e0a43f15267b4f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.893256] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Successfully created port: b401ba73-3487-46cb-81ca-b4558f83c2cb {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.896889] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c07e66-e473-4ce4-8047-c4e2d61ae7eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.904305] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf4bf65-8264-4630-9351-a1e1d2bc2dcc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.936244] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc4c059-178f-42f5-831f-79ce6e03ef94 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.944413] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e7eebb-0a24-494b-a3ec-c32c98ca623b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.958034] env[61911]: DEBUG nova.compute.provider_tree [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.977839] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.034589] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 733.099555] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.425501] env[61911]: DEBUG nova.compute.manager [req-139fa9c8-c690-4443-871e-f846e2618a17 req-fdb0b30e-f138-41f0-ad7a-751368a5d4dd service nova] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Received event network-vif-deleted-2eb47a66-a7b9-4232-982b-6b27c45817f5 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 733.461185] env[61911]: DEBUG nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.604400] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Releasing lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.604824] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 733.605028] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.605981] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f154be2d-f7e2-4631-a5a7-8dd37ba01acf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.614949] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5e66eb-9661-4d30-88c4-0fe406b0e457 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.639920] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ade5af6e-0235-407c-85e6-9668f50ccad3 could not be found. [ 733.640168] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.640351] env[61911]: INFO nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.640592] env[61911]: DEBUG oslo.service.loopingcall [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.640814] env[61911]: DEBUG nova.compute.manager [-] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 733.640909] env[61911]: DEBUG nova.network.neutron [-] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.658386] env[61911]: DEBUG nova.network.neutron [-] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.782532] env[61911]: ERROR nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. [ 733.782532] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.782532] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.782532] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.782532] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.782532] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.782532] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.782532] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.782532] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.782532] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 733.782532] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.782532] env[61911]: ERROR nova.compute.manager raise self.value [ 733.782532] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.782532] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.782532] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.782532] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.783183] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.783183] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.783183] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. [ 733.783183] env[61911]: ERROR nova.compute.manager [ 733.783183] env[61911]: Traceback (most recent call last): [ 733.783183] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.783183] env[61911]: listener.cb(fileno) [ 733.783183] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.783183] env[61911]: result = function(*args, **kwargs) [ 733.783183] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.783183] env[61911]: return func(*args, **kwargs) [ 733.783183] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 733.783183] env[61911]: raise e [ 733.783183] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.783183] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 733.783183] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.783183] env[61911]: created_port_ids = self._update_ports_for_instance( [ 733.783183] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.783183] env[61911]: with excutils.save_and_reraise_exception(): [ 733.783183] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.783183] env[61911]: self.force_reraise() [ 733.783183] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.783183] env[61911]: raise self.value [ 733.783183] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.783183] env[61911]: updated_port = self._update_port( [ 733.783183] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.783183] env[61911]: _ensure_no_port_binding_failure(port) [ 733.783183] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.783183] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.784239] env[61911]: nova.exception.PortBindingFailed: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. [ 733.784239] env[61911]: Removing descriptor: 16 [ 733.967274] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.967734] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Traceback (most recent call last): [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self.driver.spawn(context, instance, image_meta, [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] vm_ref = self.build_virtual_machine(instance, [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.967734] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] for vif in network_info: [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] return self._sync_wrapper(fn, *args, **kwargs) [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self.wait() [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self[:] = self._gt.wait() [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] return self._exit_event.wait() [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] current.throw(*self._exc) [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.968119] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] result = function(*args, **kwargs) [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] return func(*args, **kwargs) [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] raise e [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] nwinfo = self.network_api.allocate_for_instance( [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] created_port_ids = self._update_ports_for_instance( [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] with excutils.save_and_reraise_exception(): [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] self.force_reraise() [ 733.968599] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] raise self.value [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] updated_port = self._update_port( [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] _ensure_no_port_binding_failure(port) [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] raise exception.PortBindingFailed(port_id=port['id']) [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] nova.exception.PortBindingFailed: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. [ 733.969028] env[61911]: ERROR nova.compute.manager [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] [ 733.969028] env[61911]: DEBUG nova.compute.utils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 733.972046] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.034s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.972046] env[61911]: INFO nova.compute.claims [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.974333] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Build of instance e5f4224f-0002-4b75-9f69-d1ac50d955ae was re-scheduled: Binding failed for port 23386ee7-1825-4aa5-9065-66ddfcf27500, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 733.974764] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 733.974991] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.975154] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquired lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.975310] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.048121] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 734.074917] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 734.075161] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.075321] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 734.075503] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.075650] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 734.075803] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 734.076027] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 734.076202] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 734.076362] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 734.076520] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 734.076689] env[61911]: DEBUG nova.virt.hardware [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 734.077558] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409e4f26-63e9-4500-8fd6-134e82b9d0c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.085768] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3746cbf8-3e99-4bb8-8153-44285c497cb7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.099346] env[61911]: ERROR nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Traceback (most recent call last): [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] yield resources [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self.driver.spawn(context, instance, image_meta, [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] vm_ref = self.build_virtual_machine(instance, [ 734.099346] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] for vif in network_info: [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] return self._sync_wrapper(fn, *args, **kwargs) [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self.wait() [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self[:] = self._gt.wait() [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] return self._exit_event.wait() [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.099661] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] current.throw(*self._exc) [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] result = function(*args, **kwargs) [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] return func(*args, **kwargs) [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] raise e [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] nwinfo = self.network_api.allocate_for_instance( [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] created_port_ids = self._update_ports_for_instance( [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] with excutils.save_and_reraise_exception(): [ 734.100037] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self.force_reraise() [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] raise self.value [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] updated_port = self._update_port( [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] _ensure_no_port_binding_failure(port) [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] raise exception.PortBindingFailed(port_id=port['id']) [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] nova.exception.PortBindingFailed: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. [ 734.100416] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] [ 734.100416] env[61911]: INFO nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Terminating instance [ 734.160389] env[61911]: DEBUG nova.network.neutron [-] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.495444] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.582182] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.604384] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Acquiring lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.604384] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Acquired lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.604553] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.663696] env[61911]: INFO nova.compute.manager [-] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Took 1.02 seconds to deallocate network for instance. [ 734.666482] env[61911]: DEBUG nova.compute.claims [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.667514] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.085482] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Releasing lock "refresh_cache-e5f4224f-0002-4b75-9f69-d1ac50d955ae" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.085705] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 735.085878] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 735.086051] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.102976] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.124867] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.206914] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.337331] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc9d9ab-d8ec-453b-a947-5c13e5e2cf39 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.349398] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca456c0-73bb-4377-bccd-993c98bf038b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.380028] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf2dea0-1681-4035-ba8a-d44e29ca00b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.389305] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebe4848-aef2-4366-802e-33d5bfc2eada {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.401328] env[61911]: DEBUG nova.compute.provider_tree [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.472795] env[61911]: DEBUG nova.compute.manager [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Received event network-changed-b401ba73-3487-46cb-81ca-b4558f83c2cb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 735.473160] env[61911]: DEBUG nova.compute.manager [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Refreshing instance network info cache due to event network-changed-b401ba73-3487-46cb-81ca-b4558f83c2cb. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 735.473415] env[61911]: DEBUG oslo_concurrency.lockutils [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] Acquiring lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.605859] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.711678] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Releasing lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.712239] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 735.712446] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.712784] env[61911]: DEBUG oslo_concurrency.lockutils [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] Acquired lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.712985] env[61911]: DEBUG nova.network.neutron [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Refreshing network info cache for port b401ba73-3487-46cb-81ca-b4558f83c2cb {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.714299] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9504f0f-8836-4569-8fbb-8724eef28796 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.723419] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46176428-4bbf-4e58-9cc8-22907cc83989 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.746558] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7f777a69-d8ce-4a2f-83f5-085fad6f9809 could not be found. [ 735.746778] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.746955] env[61911]: INFO nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Took 0.03 seconds to destroy the instance on the hypervisor. [ 735.747238] env[61911]: DEBUG oslo.service.loopingcall [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.747472] env[61911]: DEBUG nova.compute.manager [-] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 735.747577] env[61911]: DEBUG nova.network.neutron [-] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.762924] env[61911]: DEBUG nova.network.neutron [-] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.904243] env[61911]: DEBUG nova.scheduler.client.report [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 736.108722] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: e5f4224f-0002-4b75-9f69-d1ac50d955ae] Took 1.02 seconds to deallocate network for instance. [ 736.234681] env[61911]: DEBUG nova.network.neutron [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.265177] env[61911]: DEBUG nova.network.neutron [-] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.315968] env[61911]: DEBUG nova.network.neutron [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.411211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.411211] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 736.420443] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.834s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.768270] env[61911]: INFO nova.compute.manager [-] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Took 1.02 seconds to deallocate network for instance. [ 736.770524] env[61911]: DEBUG nova.compute.claims [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.770699] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.818949] env[61911]: DEBUG oslo_concurrency.lockutils [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] Releasing lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.819161] env[61911]: DEBUG nova.compute.manager [req-4e896383-78b4-4b62-8b9e-b874fc15aef2 req-2bc86e18-356d-4b79-bcc6-732be29a3075 service nova] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Received event network-vif-deleted-b401ba73-3487-46cb-81ca-b4558f83c2cb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 736.923434] env[61911]: DEBUG nova.compute.utils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.924773] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 736.924942] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 737.127289] env[61911]: DEBUG nova.policy [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed9c79e86e4d471e96ec0f2825b4dbf5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67acac14760040fa95f06a5d080fd485', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.146434] env[61911]: INFO nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Deleted allocations for instance e5f4224f-0002-4b75-9f69-d1ac50d955ae [ 737.327310] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464ab933-e948-4d5b-8816-c406198725d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.335162] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246d8515-6645-4f8e-be5e-95e0ca31313d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.367905] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f14c5a5-b410-41a9-a66b-7eec53618d82 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.380876] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6a0b69-ea9e-41d5-880e-084891f127f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.397185] env[61911]: DEBUG nova.compute.provider_tree [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.432919] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 737.599841] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Successfully created port: f5201214-725e-4653-960a-73fb2c3ecdbb {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.661899] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "e5f4224f-0002-4b75-9f69-d1ac50d955ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.691s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.902210] env[61911]: DEBUG nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 738.165585] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 738.372441] env[61911]: DEBUG nova.compute.manager [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Received event network-changed-f5201214-725e-4653-960a-73fb2c3ecdbb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 738.372441] env[61911]: DEBUG nova.compute.manager [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Refreshing instance network info cache due to event network-changed-f5201214-725e-4653-960a-73fb2c3ecdbb. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 738.372441] env[61911]: DEBUG oslo_concurrency.lockutils [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] Acquiring lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.372441] env[61911]: DEBUG oslo_concurrency.lockutils [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] Acquired lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.372441] env[61911]: DEBUG nova.network.neutron [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Refreshing network info cache for port f5201214-725e-4653-960a-73fb2c3ecdbb {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 738.407328] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.407845] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Traceback (most recent call last): [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self.driver.spawn(context, instance, image_meta, [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] vm_ref = self.build_virtual_machine(instance, [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.407845] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] for vif in network_info: [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] return self._sync_wrapper(fn, *args, **kwargs) [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self.wait() [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self[:] = self._gt.wait() [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] return self._exit_event.wait() [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] current.throw(*self._exc) [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.408246] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] result = function(*args, **kwargs) [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] return func(*args, **kwargs) [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] raise e [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] nwinfo = self.network_api.allocate_for_instance( [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] created_port_ids = self._update_ports_for_instance( [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] with excutils.save_and_reraise_exception(): [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] self.force_reraise() [ 738.408611] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] raise self.value [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] updated_port = self._update_port( [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] _ensure_no_port_binding_failure(port) [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] raise exception.PortBindingFailed(port_id=port['id']) [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] nova.exception.PortBindingFailed: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. [ 738.408991] env[61911]: ERROR nova.compute.manager [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] [ 738.408991] env[61911]: DEBUG nova.compute.utils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.409859] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.972s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.411333] env[61911]: INFO nova.compute.claims [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.413891] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Build of instance ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325 was re-scheduled: Binding failed for port 7c19d4a7-07f8-4102-bc77-8893acbb3672, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 738.414344] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 738.414587] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.414750] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquired lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.414913] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.444022] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 738.473526] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 738.474612] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.474612] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 738.474612] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.474612] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 738.474612] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 738.475264] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 738.475789] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 738.476008] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 738.476762] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 738.476762] env[61911]: DEBUG nova.virt.hardware [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 738.478047] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7bfd61-c9c4-44d5-aeb1-a2afc5b9498e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.487717] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3937e142-1aec-41b3-8c5b-014ad78832a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.548063] env[61911]: ERROR nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. [ 738.548063] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.548063] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.548063] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.548063] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.548063] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.548063] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.548063] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.548063] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.548063] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 738.548063] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.548063] env[61911]: ERROR nova.compute.manager raise self.value [ 738.548063] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.548063] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.548063] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.548063] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.548619] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.548619] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.548619] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. [ 738.548619] env[61911]: ERROR nova.compute.manager [ 738.548619] env[61911]: Traceback (most recent call last): [ 738.548619] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.548619] env[61911]: listener.cb(fileno) [ 738.548619] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.548619] env[61911]: result = function(*args, **kwargs) [ 738.548619] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.548619] env[61911]: return func(*args, **kwargs) [ 738.548619] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 738.548619] env[61911]: raise e [ 738.548619] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.548619] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 738.548619] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.548619] env[61911]: created_port_ids = self._update_ports_for_instance( [ 738.548619] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.548619] env[61911]: with excutils.save_and_reraise_exception(): [ 738.548619] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.548619] env[61911]: self.force_reraise() [ 738.548619] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.548619] env[61911]: raise self.value [ 738.548619] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.548619] env[61911]: updated_port = self._update_port( [ 738.548619] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.548619] env[61911]: _ensure_no_port_binding_failure(port) [ 738.548619] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.548619] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.549640] env[61911]: nova.exception.PortBindingFailed: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. [ 738.549640] env[61911]: Removing descriptor: 16 [ 738.549640] env[61911]: ERROR nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Traceback (most recent call last): [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] yield resources [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self.driver.spawn(context, instance, image_meta, [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.549640] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] vm_ref = self.build_virtual_machine(instance, [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] for vif in network_info: [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return self._sync_wrapper(fn, *args, **kwargs) [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self.wait() [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self[:] = self._gt.wait() [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return self._exit_event.wait() [ 738.550221] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] result = hub.switch() [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return self.greenlet.switch() [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] result = function(*args, **kwargs) [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return func(*args, **kwargs) [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] raise e [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] nwinfo = self.network_api.allocate_for_instance( [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.550869] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] created_port_ids = self._update_ports_for_instance( [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] with excutils.save_and_reraise_exception(): [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self.force_reraise() [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] raise self.value [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] updated_port = self._update_port( [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] _ensure_no_port_binding_failure(port) [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.551615] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] raise exception.PortBindingFailed(port_id=port['id']) [ 738.552378] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] nova.exception.PortBindingFailed: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. [ 738.552378] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] [ 738.552378] env[61911]: INFO nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Terminating instance [ 738.691633] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.890960] env[61911]: DEBUG nova.network.neutron [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.944733] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.981546] env[61911]: DEBUG nova.network.neutron [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.055364] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Acquiring lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.112708] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.485719] env[61911]: DEBUG oslo_concurrency.lockutils [req-bf958749-b9c4-4cf9-9756-d055a5ef16f6 req-98cb9368-8966-4ad8-bafa-fdf0c4253005 service nova] Releasing lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.487551] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Acquired lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.488922] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.619022] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Releasing lock "refresh_cache-ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.619022] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 739.619022] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 739.619022] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.641458] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.747375] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105b47b7-37ba-4abb-8b76-5453c4e7e8c1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.754793] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f357bb5b-67f1-4084-9e23-3219029970f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.784167] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd943494-77bf-4117-b19a-3d5a94d69705 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.791480] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c586ff9c-bee1-4ce3-9155-0475d500390b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.806023] env[61911]: DEBUG nova.compute.provider_tree [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.007868] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.124545] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.143183] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.309172] env[61911]: DEBUG nova.scheduler.client.report [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 740.393534] env[61911]: DEBUG nova.compute.manager [req-8afede85-b892-454d-9986-d2aad88cf4fa req-c76e9250-307e-4b60-8738-2653d2316833 service nova] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Received event network-vif-deleted-f5201214-725e-4653-960a-73fb2c3ecdbb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 740.627324] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Releasing lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.627834] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 740.628091] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.628439] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-655c2a2c-9f0b-4289-8224-1da9b04ea6ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.637845] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623900cc-a286-4ed7-ad89-f68ebbdc66d0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.648059] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325] Took 1.03 seconds to deallocate network for instance. [ 740.660739] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea27990b-d194-4b81-b18b-37804c22ceb3 could not be found. [ 740.660938] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.661127] env[61911]: INFO nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 740.661359] env[61911]: DEBUG oslo.service.loopingcall [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.661568] env[61911]: DEBUG nova.compute.manager [-] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 740.661712] env[61911]: DEBUG nova.network.neutron [-] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.677739] env[61911]: DEBUG nova.network.neutron [-] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.813774] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.814356] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 740.817184] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.539s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.180031] env[61911]: DEBUG nova.network.neutron [-] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.322977] env[61911]: DEBUG nova.compute.utils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.323442] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 741.323604] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 741.368273] env[61911]: DEBUG nova.policy [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f585bde44cde40bcbff7f8262c4a01af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '113a8c86019d40d483097b9999561030', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.621658] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b70dd9c-becf-48bd-aff6-cae7a2b11081 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.629289] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a20e3c2-c409-40f8-a4f9-8b60836d19bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.660923] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d772f862-7ea1-40fb-8b6b-122d98a1f046 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.667983] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c1d575-a70a-471f-9276-6ab9ce6f3a12 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.681039] env[61911]: DEBUG nova.compute.provider_tree [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.686018] env[61911]: INFO nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Deleted allocations for instance ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325 [ 741.688187] env[61911]: INFO nova.compute.manager [-] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Took 1.03 seconds to deallocate network for instance. [ 741.690280] env[61911]: DEBUG nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 741.695751] env[61911]: DEBUG nova.compute.claims [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.695751] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.830940] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 741.880522] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Successfully created port: 69ab0760-59b8-45bf-9955-8905b47a9ed1 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.199019] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "ee6d3ffa-5fbb-40cf-9b68-05b0c80b5325" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.194s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.199019] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.380s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.199019] env[61911]: ERROR nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. [ 742.199019] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Traceback (most recent call last): [ 742.199019] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 742.199019] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self.driver.spawn(context, instance, image_meta, [ 742.199019] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] vm_ref = self.build_virtual_machine(instance, [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] for vif in network_info: [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] return self._sync_wrapper(fn, *args, **kwargs) [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self.wait() [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.199658] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self[:] = self._gt.wait() [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] return self._exit_event.wait() [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] current.throw(*self._exc) [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] result = function(*args, **kwargs) [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] return func(*args, **kwargs) [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] raise e [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] nwinfo = self.network_api.allocate_for_instance( [ 742.200123] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] created_port_ids = self._update_ports_for_instance( [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] with excutils.save_and_reraise_exception(): [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] self.force_reraise() [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] raise self.value [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] updated_port = self._update_port( [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] _ensure_no_port_binding_failure(port) [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.200534] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] raise exception.PortBindingFailed(port_id=port['id']) [ 742.200933] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] nova.exception.PortBindingFailed: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. [ 742.200933] env[61911]: ERROR nova.compute.manager [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] [ 742.200933] env[61911]: DEBUG nova.compute.utils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.202908] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.356s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.204543] env[61911]: INFO nova.compute.claims [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.207848] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Build of instance a09bc118-c3f8-4943-89a1-8a0c59d528c5 was re-scheduled: Binding failed for port 8486f934-8057-4b47-9f6b-4449b35edb4e, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 742.210616] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 742.210616] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquiring lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.210616] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Acquired lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.210616] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.711940] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 742.737434] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.762063] env[61911]: DEBUG nova.compute.manager [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Received event network-changed-69ab0760-59b8-45bf-9955-8905b47a9ed1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 742.762063] env[61911]: DEBUG nova.compute.manager [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Refreshing instance network info cache due to event network-changed-69ab0760-59b8-45bf-9955-8905b47a9ed1. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 742.762063] env[61911]: DEBUG oslo_concurrency.lockutils [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] Acquiring lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.762063] env[61911]: DEBUG oslo_concurrency.lockutils [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] Acquired lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.762063] env[61911]: DEBUG nova.network.neutron [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Refreshing network info cache for port 69ab0760-59b8-45bf-9955-8905b47a9ed1 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 742.825934] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.843602] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 742.868116] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 742.868375] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.868534] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 742.868717] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.868861] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 742.869030] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 742.869240] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 742.869416] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 742.869618] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 742.869739] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 742.869910] env[61911]: DEBUG nova.virt.hardware [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 742.870805] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f748a060-88b6-4d0b-9b44-0fe3e51fe2b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.879289] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b73e924-6e11-4391-a2db-5d2ca0e11227 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.948305] env[61911]: ERROR nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. [ 742.948305] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 742.948305] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.948305] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 742.948305] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.948305] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 742.948305] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.948305] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 742.948305] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.948305] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 742.948305] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.948305] env[61911]: ERROR nova.compute.manager raise self.value [ 742.948305] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.948305] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 742.948305] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.948305] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 742.948781] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.948781] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 742.948781] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. [ 742.948781] env[61911]: ERROR nova.compute.manager [ 742.948781] env[61911]: Traceback (most recent call last): [ 742.948781] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 742.948781] env[61911]: listener.cb(fileno) [ 742.948781] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.948781] env[61911]: result = function(*args, **kwargs) [ 742.948781] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.948781] env[61911]: return func(*args, **kwargs) [ 742.948781] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.948781] env[61911]: raise e [ 742.948781] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.948781] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 742.948781] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.948781] env[61911]: created_port_ids = self._update_ports_for_instance( [ 742.948781] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.948781] env[61911]: with excutils.save_and_reraise_exception(): [ 742.948781] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.948781] env[61911]: self.force_reraise() [ 742.948781] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.948781] env[61911]: raise self.value [ 742.948781] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.948781] env[61911]: updated_port = self._update_port( [ 742.948781] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.948781] env[61911]: _ensure_no_port_binding_failure(port) [ 742.948781] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.948781] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 742.949666] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. [ 742.949666] env[61911]: Removing descriptor: 19 [ 742.949666] env[61911]: ERROR nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Traceback (most recent call last): [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] yield resources [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self.driver.spawn(context, instance, image_meta, [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.949666] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] vm_ref = self.build_virtual_machine(instance, [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] for vif in network_info: [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return self._sync_wrapper(fn, *args, **kwargs) [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self.wait() [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self[:] = self._gt.wait() [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return self._exit_event.wait() [ 742.950137] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] result = hub.switch() [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return self.greenlet.switch() [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] result = function(*args, **kwargs) [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return func(*args, **kwargs) [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] raise e [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] nwinfo = self.network_api.allocate_for_instance( [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.950602] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] created_port_ids = self._update_ports_for_instance( [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] with excutils.save_and_reraise_exception(): [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self.force_reraise() [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] raise self.value [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] updated_port = self._update_port( [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] _ensure_no_port_binding_failure(port) [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.951022] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] raise exception.PortBindingFailed(port_id=port['id']) [ 742.951417] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] nova.exception.PortBindingFailed: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. [ 742.951417] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] [ 742.951417] env[61911]: INFO nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Terminating instance [ 743.236592] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.289296] env[61911]: DEBUG nova.network.neutron [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.330997] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Releasing lock "refresh_cache-a09bc118-c3f8-4943-89a1-8a0c59d528c5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.331247] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 743.331425] env[61911]: DEBUG nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 743.331584] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.346086] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.389891] env[61911]: DEBUG nova.network.neutron [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.456631] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.500013] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973c71f0-e24a-4a79-9559-4ae9e8549802 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.507686] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925d8095-da89-4c46-bdf5-e7f3e6bc0d6c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.539589] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfd2979-a67a-42fd-8fce-74ca8d978d98 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.550147] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b31c7ea-9e3d-49c5-88ec-ce32a1f53835 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.566449] env[61911]: DEBUG nova.compute.provider_tree [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.849069] env[61911]: DEBUG nova.network.neutron [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.893025] env[61911]: DEBUG oslo_concurrency.lockutils [req-b3c35fa2-2fa7-4d03-871a-c80f58d0f11b req-292ecd68-3c5d-4637-ad94-4d180203e551 service nova] Releasing lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.893025] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquired lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.893025] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.071322] env[61911]: DEBUG nova.scheduler.client.report [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 744.352360] env[61911]: INFO nova.compute.manager [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] [instance: a09bc118-c3f8-4943-89a1-8a0c59d528c5] Took 1.02 seconds to deallocate network for instance. [ 744.409895] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.478148] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.576233] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.576769] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 744.579944] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.716s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.582172] env[61911]: INFO nova.compute.claims [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.787392] env[61911]: DEBUG nova.compute.manager [req-4cda4809-8dec-4de8-a71d-ed0094435332 req-9461e21d-41d6-4ccf-ade4-c534887141df service nova] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Received event network-vif-deleted-69ab0760-59b8-45bf-9955-8905b47a9ed1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 744.980983] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Releasing lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.981469] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 744.981668] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.981969] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fffbd5da-1c17-4585-b71b-20d82c23ce99 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.991012] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6669c24-7317-4e1c-8b6c-b653822ecc3d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.015714] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9 could not be found. [ 745.015714] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.015714] env[61911]: INFO nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 745.015714] env[61911]: DEBUG oslo.service.loopingcall [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.015714] env[61911]: DEBUG nova.compute.manager [-] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 745.015714] env[61911]: DEBUG nova.network.neutron [-] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.029893] env[61911]: DEBUG nova.network.neutron [-] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.090038] env[61911]: DEBUG nova.compute.utils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.091327] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 745.091327] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.132899] env[61911]: DEBUG nova.policy [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c323c45a848419fa3ed8bb189bbbeae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e9c9e86361242598a0fe512d1cb65c5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.388454] env[61911]: INFO nova.scheduler.client.report [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Deleted allocations for instance a09bc118-c3f8-4943-89a1-8a0c59d528c5 [ 745.508763] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Successfully created port: d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.532355] env[61911]: DEBUG nova.network.neutron [-] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.595333] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 745.899782] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f468819e-0e6c-468d-ad84-111cd0ac319e tempest-ListServersNegativeTestJSON-531723987 tempest-ListServersNegativeTestJSON-531723987-project-member] Lock "a09bc118-c3f8-4943-89a1-8a0c59d528c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.861s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.904775] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6aa669-a26c-4d98-9390-71078cfce05b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.910199] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571e60fa-e4f2-41f8-849e-babb56acf6e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.940308] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e49d557-ac2e-4af0-88ef-0db364eb0cc2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.947824] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1a1873-8d26-477b-b472-77c5de670440 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.964343] env[61911]: DEBUG nova.compute.provider_tree [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.034865] env[61911]: INFO nova.compute.manager [-] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Took 1.02 seconds to deallocate network for instance. [ 746.037209] env[61911]: DEBUG nova.compute.claims [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 746.037363] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.406273] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 746.467407] env[61911]: DEBUG nova.scheduler.client.report [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 746.607734] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 746.633283] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 746.634201] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.634409] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 746.634612] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.634761] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 746.634909] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 746.635134] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 746.635298] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 746.635462] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 746.635630] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 746.635803] env[61911]: DEBUG nova.virt.hardware [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 746.636747] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7f26dc-4cdd-4f7d-9612-f4b0469c9244 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.644618] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05be9bdd-a3f3-40f6-b481-9da4f01ba7b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.665771] env[61911]: ERROR nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. [ 746.665771] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.665771] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 746.665771] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.665771] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.665771] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.665771] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.665771] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.665771] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.665771] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 746.665771] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.665771] env[61911]: ERROR nova.compute.manager raise self.value [ 746.665771] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.665771] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.665771] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.665771] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.666266] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.666266] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.666266] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. [ 746.666266] env[61911]: ERROR nova.compute.manager [ 746.666266] env[61911]: Traceback (most recent call last): [ 746.666266] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.666266] env[61911]: listener.cb(fileno) [ 746.666266] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.666266] env[61911]: result = function(*args, **kwargs) [ 746.666266] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.666266] env[61911]: return func(*args, **kwargs) [ 746.666266] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 746.666266] env[61911]: raise e [ 746.666266] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 746.666266] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 746.666266] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.666266] env[61911]: created_port_ids = self._update_ports_for_instance( [ 746.666266] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.666266] env[61911]: with excutils.save_and_reraise_exception(): [ 746.666266] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.666266] env[61911]: self.force_reraise() [ 746.666266] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.666266] env[61911]: raise self.value [ 746.666266] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.666266] env[61911]: updated_port = self._update_port( [ 746.666266] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.666266] env[61911]: _ensure_no_port_binding_failure(port) [ 746.666266] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.666266] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.667257] env[61911]: nova.exception.PortBindingFailed: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. [ 746.667257] env[61911]: Removing descriptor: 19 [ 746.667257] env[61911]: ERROR nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Traceback (most recent call last): [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] yield resources [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self.driver.spawn(context, instance, image_meta, [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.667257] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] vm_ref = self.build_virtual_machine(instance, [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] for vif in network_info: [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return self._sync_wrapper(fn, *args, **kwargs) [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self.wait() [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self[:] = self._gt.wait() [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return self._exit_event.wait() [ 746.667602] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] result = hub.switch() [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return self.greenlet.switch() [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] result = function(*args, **kwargs) [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return func(*args, **kwargs) [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] raise e [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] nwinfo = self.network_api.allocate_for_instance( [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.667968] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] created_port_ids = self._update_ports_for_instance( [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] with excutils.save_and_reraise_exception(): [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self.force_reraise() [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] raise self.value [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] updated_port = self._update_port( [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] _ensure_no_port_binding_failure(port) [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.668350] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] raise exception.PortBindingFailed(port_id=port['id']) [ 746.668698] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] nova.exception.PortBindingFailed: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. [ 746.668698] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] [ 746.668698] env[61911]: INFO nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Terminating instance [ 746.811140] env[61911]: DEBUG nova.compute.manager [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Received event network-changed-d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 746.811235] env[61911]: DEBUG nova.compute.manager [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Refreshing instance network info cache due to event network-changed-d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 746.811460] env[61911]: DEBUG oslo_concurrency.lockutils [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] Acquiring lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.811604] env[61911]: DEBUG oslo_concurrency.lockutils [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] Acquired lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.811767] env[61911]: DEBUG nova.network.neutron [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Refreshing network info cache for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 746.927618] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.972700] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.973276] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 746.976136] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.226s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.977621] env[61911]: INFO nova.compute.claims [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.172085] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Acquiring lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.331044] env[61911]: DEBUG nova.network.neutron [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.460340] env[61911]: DEBUG nova.network.neutron [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.481665] env[61911]: DEBUG nova.compute.utils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.484787] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 747.484955] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.577484] env[61911]: DEBUG nova.policy [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4dcd5ff07d344d65b88f88d4f16744d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec782daef4cf4deeaa47936b07b41dc3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.964596] env[61911]: DEBUG oslo_concurrency.lockutils [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] Releasing lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.964894] env[61911]: DEBUG nova.compute.manager [req-3fd02cce-2480-4c33-b644-0f4902b7cee7 req-d073b7c6-d606-4261-905c-324521ca6f21 service nova] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Received event network-vif-deleted-d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 747.965284] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Acquired lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.965514] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.985297] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 748.305621] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Successfully created port: a9416ca8-d180-4e95-a8d7-83834df2d258 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.340961] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831e7ca6-9747-4f83-93da-5455a7ac07ad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.349094] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d386e4-1ab2-45fe-b2bd-0730df991f13 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.382705] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9c90a1-276c-444b-b3aa-5a2358931d59 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.394251] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbe7191-df55-47ae-891a-2a758b9307e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.409269] env[61911]: DEBUG nova.compute.provider_tree [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.484405] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.573313] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.912334] env[61911]: DEBUG nova.scheduler.client.report [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 749.000246] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 749.029721] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 749.029982] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.030167] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 749.030380] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.030528] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 749.030676] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 749.030877] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 749.031064] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 749.031240] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 749.031405] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 749.031573] env[61911]: DEBUG nova.virt.hardware [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 749.032466] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf70459a-fd8a-4fd3-b3f1-3efa574e7e3e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.040891] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf50d88-2cc4-4871-9843-8f465d60f72a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.076927] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Releasing lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.077394] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 749.077607] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 749.078071] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-936fa86f-ff0b-407b-b205-df6d8922c206 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.088980] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7260997-547e-4d35-ab4b-5fc223e714bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.111821] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe46e0e7-c1aa-4bf5-af01-cdb04031f945 could not be found. [ 749.112122] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.112340] env[61911]: INFO nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Took 0.03 seconds to destroy the instance on the hypervisor. [ 749.112593] env[61911]: DEBUG oslo.service.loopingcall [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.112813] env[61911]: DEBUG nova.compute.manager [-] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 749.112932] env[61911]: DEBUG nova.network.neutron [-] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.134191] env[61911]: DEBUG nova.network.neutron [-] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.179523] env[61911]: DEBUG nova.compute.manager [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Received event network-changed-a9416ca8-d180-4e95-a8d7-83834df2d258 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 749.179710] env[61911]: DEBUG nova.compute.manager [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Refreshing instance network info cache due to event network-changed-a9416ca8-d180-4e95-a8d7-83834df2d258. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 749.179921] env[61911]: DEBUG oslo_concurrency.lockutils [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] Acquiring lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.180075] env[61911]: DEBUG oslo_concurrency.lockutils [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] Acquired lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.180288] env[61911]: DEBUG nova.network.neutron [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Refreshing network info cache for port a9416ca8-d180-4e95-a8d7-83834df2d258 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.391661] env[61911]: ERROR nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. [ 749.391661] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.391661] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.391661] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.391661] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.391661] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.391661] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.391661] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.391661] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.391661] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 749.391661] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.391661] env[61911]: ERROR nova.compute.manager raise self.value [ 749.391661] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.391661] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.391661] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.391661] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.392125] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.392125] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.392125] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. [ 749.392125] env[61911]: ERROR nova.compute.manager [ 749.392125] env[61911]: Traceback (most recent call last): [ 749.392125] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.392125] env[61911]: listener.cb(fileno) [ 749.392125] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.392125] env[61911]: result = function(*args, **kwargs) [ 749.392125] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.392125] env[61911]: return func(*args, **kwargs) [ 749.392125] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 749.392125] env[61911]: raise e [ 749.392125] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.392125] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 749.392125] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.392125] env[61911]: created_port_ids = self._update_ports_for_instance( [ 749.392125] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.392125] env[61911]: with excutils.save_and_reraise_exception(): [ 749.392125] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.392125] env[61911]: self.force_reraise() [ 749.392125] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.392125] env[61911]: raise self.value [ 749.392125] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.392125] env[61911]: updated_port = self._update_port( [ 749.392125] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.392125] env[61911]: _ensure_no_port_binding_failure(port) [ 749.392125] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.392125] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.392984] env[61911]: nova.exception.PortBindingFailed: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. [ 749.392984] env[61911]: Removing descriptor: 19 [ 749.392984] env[61911]: ERROR nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Traceback (most recent call last): [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] yield resources [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self.driver.spawn(context, instance, image_meta, [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.392984] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] vm_ref = self.build_virtual_machine(instance, [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] for vif in network_info: [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return self._sync_wrapper(fn, *args, **kwargs) [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self.wait() [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self[:] = self._gt.wait() [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return self._exit_event.wait() [ 749.393336] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] result = hub.switch() [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return self.greenlet.switch() [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] result = function(*args, **kwargs) [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return func(*args, **kwargs) [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] raise e [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] nwinfo = self.network_api.allocate_for_instance( [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.393908] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] created_port_ids = self._update_ports_for_instance( [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] with excutils.save_and_reraise_exception(): [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self.force_reraise() [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] raise self.value [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] updated_port = self._update_port( [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] _ensure_no_port_binding_failure(port) [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.394595] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] raise exception.PortBindingFailed(port_id=port['id']) [ 749.395093] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] nova.exception.PortBindingFailed: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. [ 749.395093] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] [ 749.395093] env[61911]: INFO nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Terminating instance [ 749.417089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.417643] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 749.422639] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.754s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.636479] env[61911]: DEBUG nova.network.neutron [-] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.698036] env[61911]: DEBUG nova.network.neutron [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.797444] env[61911]: DEBUG nova.network.neutron [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.897557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Acquiring lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.925312] env[61911]: DEBUG nova.compute.utils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.930019] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 749.930019] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 749.967893] env[61911]: DEBUG nova.policy [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f987f2f2bb29466b81ab84f346e179d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a427d184a154ae0bc1e9d061633c360', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.142506] env[61911]: INFO nova.compute.manager [-] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Took 1.03 seconds to deallocate network for instance. [ 750.147567] env[61911]: DEBUG nova.compute.claims [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 750.147766] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.233856] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999b6ff2-dfba-49fb-8310-25448a9441e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.241985] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b4f0c9-b3db-45fa-a6fd-27293f78a569 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.280238] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3405bfa-767b-4aa5-82c5-71e3b0704cb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.288527] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcee3f0-beff-40e8-bc66-f612887e35ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.305056] env[61911]: DEBUG oslo_concurrency.lockutils [req-42caeb92-0087-4b35-ba29-e9bf6bbb5603 req-e50c6231-7a57-44c8-ab24-03b909c2b19f service nova] Releasing lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.305873] env[61911]: DEBUG nova.compute.provider_tree [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.307516] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Acquired lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.307683] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.319581] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Successfully created port: c086fba1-2024-4244-b3aa-c7e3c3b470ae {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.432720] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 750.812089] env[61911]: DEBUG nova.scheduler.client.report [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 750.831401] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.917569] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.212583] env[61911]: DEBUG nova.compute.manager [req-a245d6f8-6ad0-4714-8827-5713ae9cf098 req-43efeaba-ab32-4080-a879-41ad3bab9dc8 service nova] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Received event network-vif-deleted-a9416ca8-d180-4e95-a8d7-83834df2d258 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 751.317657] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.319036] env[61911]: ERROR nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Traceback (most recent call last): [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self.driver.spawn(context, instance, image_meta, [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] vm_ref = self.build_virtual_machine(instance, [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.319036] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] for vif in network_info: [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] return self._sync_wrapper(fn, *args, **kwargs) [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self.wait() [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self[:] = self._gt.wait() [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] return self._exit_event.wait() [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] current.throw(*self._exc) [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.319515] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] result = function(*args, **kwargs) [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] return func(*args, **kwargs) [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] raise e [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] nwinfo = self.network_api.allocate_for_instance( [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] created_port_ids = self._update_ports_for_instance( [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] with excutils.save_and_reraise_exception(): [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] self.force_reraise() [ 751.320060] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] raise self.value [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] updated_port = self._update_port( [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] _ensure_no_port_binding_failure(port) [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] raise exception.PortBindingFailed(port_id=port['id']) [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] nova.exception.PortBindingFailed: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. [ 751.320553] env[61911]: ERROR nova.compute.manager [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] [ 751.321469] env[61911]: DEBUG nova.compute.utils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.323279] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.553s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.332718] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Build of instance ade5af6e-0235-407c-85e6-9668f50ccad3 was re-scheduled: Binding failed for port 2eb47a66-a7b9-4232-982b-6b27c45817f5, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 751.332718] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 751.333044] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Acquiring lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.333271] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Acquired lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.333517] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.420362] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Releasing lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.420612] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 751.420814] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.421478] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54467de1-11dc-42f2-b9a5-d226210ad7a5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.430659] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebb6595-bfac-44fb-88ab-04134e8b622e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.442870] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 751.455399] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6773f2b-df86-4a4c-9b20-a314e81a7aa0 could not be found. [ 751.455568] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.455749] env[61911]: INFO nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 751.455988] env[61911]: DEBUG oslo.service.loopingcall [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.456225] env[61911]: DEBUG nova.compute.manager [-] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 751.456318] env[61911]: DEBUG nova.network.neutron [-] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.478397] env[61911]: DEBUG nova.network.neutron [-] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.484997] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 751.485258] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.485471] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 751.485589] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.485737] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 751.485926] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 751.489336] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 751.489518] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 751.489726] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 751.489945] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 751.490146] env[61911]: DEBUG nova.virt.hardware [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 751.491017] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f088c86d-93e0-46cb-8701-aec5873974ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.494867] env[61911]: DEBUG nova.network.neutron [-] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.506196] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c70f87a-db38-452c-bf1a-bef721e21631 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.662413] env[61911]: ERROR nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. [ 751.662413] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 751.662413] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 751.662413] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 751.662413] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.662413] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 751.662413] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.662413] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 751.662413] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.662413] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 751.662413] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.662413] env[61911]: ERROR nova.compute.manager raise self.value [ 751.662413] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.662413] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 751.662413] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.662413] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 751.662942] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.662942] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 751.662942] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. [ 751.662942] env[61911]: ERROR nova.compute.manager [ 751.662942] env[61911]: Traceback (most recent call last): [ 751.662942] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 751.662942] env[61911]: listener.cb(fileno) [ 751.662942] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.662942] env[61911]: result = function(*args, **kwargs) [ 751.662942] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.662942] env[61911]: return func(*args, **kwargs) [ 751.662942] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 751.662942] env[61911]: raise e [ 751.662942] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 751.662942] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 751.662942] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.662942] env[61911]: created_port_ids = self._update_ports_for_instance( [ 751.662942] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.662942] env[61911]: with excutils.save_and_reraise_exception(): [ 751.662942] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.662942] env[61911]: self.force_reraise() [ 751.662942] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.662942] env[61911]: raise self.value [ 751.662942] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.662942] env[61911]: updated_port = self._update_port( [ 751.662942] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.662942] env[61911]: _ensure_no_port_binding_failure(port) [ 751.662942] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.662942] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 751.663887] env[61911]: nova.exception.PortBindingFailed: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. [ 751.663887] env[61911]: Removing descriptor: 19 [ 751.663887] env[61911]: ERROR nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Traceback (most recent call last): [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] yield resources [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self.driver.spawn(context, instance, image_meta, [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.663887] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] vm_ref = self.build_virtual_machine(instance, [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] for vif in network_info: [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return self._sync_wrapper(fn, *args, **kwargs) [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self.wait() [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self[:] = self._gt.wait() [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return self._exit_event.wait() [ 751.664469] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] result = hub.switch() [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return self.greenlet.switch() [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] result = function(*args, **kwargs) [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return func(*args, **kwargs) [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] raise e [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] nwinfo = self.network_api.allocate_for_instance( [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.664906] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] created_port_ids = self._update_ports_for_instance( [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] with excutils.save_and_reraise_exception(): [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self.force_reraise() [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] raise self.value [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] updated_port = self._update_port( [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] _ensure_no_port_binding_failure(port) [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.665560] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] raise exception.PortBindingFailed(port_id=port['id']) [ 751.665986] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] nova.exception.PortBindingFailed: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. [ 751.665986] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] [ 751.665986] env[61911]: INFO nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Terminating instance [ 751.856214] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.985423] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.001112] env[61911]: INFO nova.compute.manager [-] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Took 0.54 seconds to deallocate network for instance. [ 752.004129] env[61911]: DEBUG nova.compute.claims [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.004338] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.072023] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.072253] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.156300] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c354f525-316f-48c3-baa1-f718a9ccfc31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.163997] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1625ca-5c2e-4d6a-bb63-e281643b5530 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.169662] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Acquiring lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.169852] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Acquired lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.170037] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 752.196623] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b34e83-904b-49df-a610-f218d6bdbef6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.204682] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b877be-e26b-4939-8554-0dc243484997 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.218716] env[61911]: DEBUG nova.compute.provider_tree [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.490321] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Releasing lock "refresh_cache-ade5af6e-0235-407c-85e6-9668f50ccad3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.490511] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 752.490699] env[61911]: DEBUG nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 752.490861] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.513071] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.721103] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.724083] env[61911]: DEBUG nova.scheduler.client.report [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 752.833881] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.015834] env[61911]: DEBUG nova.network.neutron [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.229726] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.230383] env[61911]: ERROR nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Traceback (most recent call last): [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self.driver.spawn(context, instance, image_meta, [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] vm_ref = self.build_virtual_machine(instance, [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.230383] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] for vif in network_info: [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] return self._sync_wrapper(fn, *args, **kwargs) [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self.wait() [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self[:] = self._gt.wait() [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] return self._exit_event.wait() [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] current.throw(*self._exc) [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.230814] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] result = function(*args, **kwargs) [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] return func(*args, **kwargs) [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] raise e [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] nwinfo = self.network_api.allocate_for_instance( [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] created_port_ids = self._update_ports_for_instance( [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] with excutils.save_and_reraise_exception(): [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] self.force_reraise() [ 753.231243] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] raise self.value [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] updated_port = self._update_port( [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] _ensure_no_port_binding_failure(port) [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] raise exception.PortBindingFailed(port_id=port['id']) [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] nova.exception.PortBindingFailed: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. [ 753.231660] env[61911]: ERROR nova.compute.manager [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] [ 753.231660] env[61911]: DEBUG nova.compute.utils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.232338] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.541s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.233810] env[61911]: INFO nova.compute.claims [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.239187] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Build of instance 7f777a69-d8ce-4a2f-83f5-085fad6f9809 was re-scheduled: Binding failed for port b401ba73-3487-46cb-81ca-b4558f83c2cb, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 753.239606] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 753.239821] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Acquiring lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.239958] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Acquired lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.240126] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.300583] env[61911]: DEBUG nova.compute.manager [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Received event network-changed-c086fba1-2024-4244-b3aa-c7e3c3b470ae {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 753.300825] env[61911]: DEBUG nova.compute.manager [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Refreshing instance network info cache due to event network-changed-c086fba1-2024-4244-b3aa-c7e3c3b470ae. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 753.301027] env[61911]: DEBUG oslo_concurrency.lockutils [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] Acquiring lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.336919] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Releasing lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.337431] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 753.337669] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 753.338012] env[61911]: DEBUG oslo_concurrency.lockutils [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] Acquired lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.338248] env[61911]: DEBUG nova.network.neutron [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Refreshing network info cache for port c086fba1-2024-4244-b3aa-c7e3c3b470ae {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.339307] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-407f6adc-e403-4806-bd9e-75e4ed2da056 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.348766] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1f5a86-11c8-44d4-8ead-5368a0eb7422 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.370961] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b712ffa7-19d8-47d8-8d60-48ca52735776 could not be found. [ 753.371081] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 753.372024] env[61911]: INFO nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Took 0.03 seconds to destroy the instance on the hypervisor. [ 753.372024] env[61911]: DEBUG oslo.service.loopingcall [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.372024] env[61911]: DEBUG nova.compute.manager [-] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 753.372024] env[61911]: DEBUG nova.network.neutron [-] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 753.390700] env[61911]: DEBUG nova.network.neutron [-] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.518656] env[61911]: INFO nova.compute.manager [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] [instance: ade5af6e-0235-407c-85e6-9668f50ccad3] Took 1.03 seconds to deallocate network for instance. [ 753.767082] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.859154] env[61911]: DEBUG nova.network.neutron [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.869397] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.893129] env[61911]: DEBUG nova.network.neutron [-] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.928383] env[61911]: DEBUG nova.network.neutron [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.374102] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Releasing lock "refresh_cache-7f777a69-d8ce-4a2f-83f5-085fad6f9809" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.374372] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 754.374589] env[61911]: DEBUG nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 754.374806] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.391734] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.397501] env[61911]: INFO nova.compute.manager [-] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Took 1.03 seconds to deallocate network for instance. [ 754.399472] env[61911]: DEBUG nova.compute.claims [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 754.399641] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.431284] env[61911]: DEBUG oslo_concurrency.lockutils [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] Releasing lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.431572] env[61911]: DEBUG nova.compute.manager [req-496c1152-1de6-4502-9f62-296383bb9835 req-700631da-bf43-46e8-9f4d-fe6e08256aa6 service nova] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Received event network-vif-deleted-c086fba1-2024-4244-b3aa-c7e3c3b470ae {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 754.535156] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4af891-0522-4ab2-a3ed-ad5a33a6d4cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.543098] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0623b873-3883-473a-be8c-167767fdcaa0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.546720] env[61911]: INFO nova.scheduler.client.report [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Deleted allocations for instance ade5af6e-0235-407c-85e6-9668f50ccad3 [ 754.581591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f329ce-e0d8-4ada-a28a-93bbb1cc7dad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.589873] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd97864-bbe1-486e-89a6-68f1b69fdb40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.605387] env[61911]: DEBUG nova.compute.provider_tree [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.894938] env[61911]: DEBUG nova.network.neutron [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.081796] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3710f77-9727-4e98-afb2-f0de38d1c141 tempest-ServersTestFqdnHostnames-114920031 tempest-ServersTestFqdnHostnames-114920031-project-member] Lock "ade5af6e-0235-407c-85e6-9668f50ccad3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.371s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.109605] env[61911]: DEBUG nova.scheduler.client.report [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 755.397125] env[61911]: INFO nova.compute.manager [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] [instance: 7f777a69-d8ce-4a2f-83f5-085fad6f9809] Took 1.02 seconds to deallocate network for instance. [ 755.584674] env[61911]: DEBUG nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 755.614701] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.615250] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 755.618220] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.923s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.106316] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.122781] env[61911]: DEBUG nova.compute.utils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.127434] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 756.127610] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.185905] env[61911]: DEBUG nova.policy [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c57bc181d3f746d39b8beb5e8aae3839', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4806147e49484ae2887767632c518e6d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.414304] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd679b7e-9c69-4a71-a504-0388bbc4489d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.424678] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a651f8-a78b-44e8-bd26-e4ba0cc06709 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.458113] env[61911]: INFO nova.scheduler.client.report [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Deleted allocations for instance 7f777a69-d8ce-4a2f-83f5-085fad6f9809 [ 756.463651] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb759538-3738-4449-8a84-6ca9fd7a8cbe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.472417] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643c3524-e5d8-402b-8df0-50791ac6ad21 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.488267] env[61911]: DEBUG nova.compute.provider_tree [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.630795] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 756.780355] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Successfully created port: d8b0c98e-b383-4175-a292-5a81ca0174a1 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.971229] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f92e0e58-d045-4cc4-93f7-87a357b742f4 tempest-ServerActionsTestOtherA-1263598269 tempest-ServerActionsTestOtherA-1263598269-project-member] Lock "7f777a69-d8ce-4a2f-83f5-085fad6f9809" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.674s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.991288] env[61911]: DEBUG nova.scheduler.client.report [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 757.476616] env[61911]: DEBUG nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 757.497528] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.879s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.499219] env[61911]: ERROR nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Traceback (most recent call last): [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self.driver.spawn(context, instance, image_meta, [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] vm_ref = self.build_virtual_machine(instance, [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.499219] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] for vif in network_info: [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return self._sync_wrapper(fn, *args, **kwargs) [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self.wait() [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self[:] = self._gt.wait() [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return self._exit_event.wait() [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] result = hub.switch() [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 757.499596] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return self.greenlet.switch() [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] result = function(*args, **kwargs) [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] return func(*args, **kwargs) [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] raise e [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] nwinfo = self.network_api.allocate_for_instance( [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] created_port_ids = self._update_ports_for_instance( [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] with excutils.save_and_reraise_exception(): [ 757.499990] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] self.force_reraise() [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] raise self.value [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] updated_port = self._update_port( [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] _ensure_no_port_binding_failure(port) [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] raise exception.PortBindingFailed(port_id=port['id']) [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] nova.exception.PortBindingFailed: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. [ 757.500393] env[61911]: ERROR nova.compute.manager [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] [ 757.501265] env[61911]: DEBUG nova.compute.utils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.503142] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.267s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.506029] env[61911]: INFO nova.compute.claims [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.509404] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Build of instance ea27990b-d194-4b81-b18b-37804c22ceb3 was re-scheduled: Binding failed for port f5201214-725e-4653-960a-73fb2c3ecdbb, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 757.509935] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 757.510379] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Acquiring lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.510714] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Acquired lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.511871] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.640802] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 757.670019] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 757.670245] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.670324] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 757.670538] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.670700] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 757.671218] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 757.671455] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 757.671650] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 757.671990] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 757.672240] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 757.672992] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 757.674135] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d5049c-df41-46fe-ae24-80d7805b8ca1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.683017] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0fd89b-0fdd-4e0c-86f3-65f7a9ce0311 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.005948] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.038221] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.126113] env[61911]: DEBUG nova.compute.manager [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Received event network-changed-d8b0c98e-b383-4175-a292-5a81ca0174a1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 758.126113] env[61911]: DEBUG nova.compute.manager [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Refreshing instance network info cache due to event network-changed-d8b0c98e-b383-4175-a292-5a81ca0174a1. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 758.126696] env[61911]: DEBUG oslo_concurrency.lockutils [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] Acquiring lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.126899] env[61911]: DEBUG oslo_concurrency.lockutils [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] Acquired lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.127081] env[61911]: DEBUG nova.network.neutron [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Refreshing network info cache for port d8b0c98e-b383-4175-a292-5a81ca0174a1 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.211573] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.460463] env[61911]: ERROR nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. [ 758.460463] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.460463] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 758.460463] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.460463] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.460463] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.460463] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.460463] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.460463] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.460463] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 758.460463] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.460463] env[61911]: ERROR nova.compute.manager raise self.value [ 758.460463] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.460463] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.460463] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.460463] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.461067] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.461067] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.461067] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. [ 758.461067] env[61911]: ERROR nova.compute.manager [ 758.461067] env[61911]: Traceback (most recent call last): [ 758.461067] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.461067] env[61911]: listener.cb(fileno) [ 758.461067] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.461067] env[61911]: result = function(*args, **kwargs) [ 758.461067] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.461067] env[61911]: return func(*args, **kwargs) [ 758.461067] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 758.461067] env[61911]: raise e [ 758.461067] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 758.461067] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 758.461067] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.461067] env[61911]: created_port_ids = self._update_ports_for_instance( [ 758.461067] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.461067] env[61911]: with excutils.save_and_reraise_exception(): [ 758.461067] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.461067] env[61911]: self.force_reraise() [ 758.461067] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.461067] env[61911]: raise self.value [ 758.461067] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.461067] env[61911]: updated_port = self._update_port( [ 758.461067] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.461067] env[61911]: _ensure_no_port_binding_failure(port) [ 758.461067] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.461067] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.461996] env[61911]: nova.exception.PortBindingFailed: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. [ 758.461996] env[61911]: Removing descriptor: 16 [ 758.461996] env[61911]: ERROR nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Traceback (most recent call last): [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] yield resources [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self.driver.spawn(context, instance, image_meta, [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.461996] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] vm_ref = self.build_virtual_machine(instance, [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] for vif in network_info: [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return self._sync_wrapper(fn, *args, **kwargs) [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self.wait() [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self[:] = self._gt.wait() [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return self._exit_event.wait() [ 758.462390] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] result = hub.switch() [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return self.greenlet.switch() [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] result = function(*args, **kwargs) [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return func(*args, **kwargs) [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] raise e [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] nwinfo = self.network_api.allocate_for_instance( [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.462827] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] created_port_ids = self._update_ports_for_instance( [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] with excutils.save_and_reraise_exception(): [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self.force_reraise() [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] raise self.value [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] updated_port = self._update_port( [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] _ensure_no_port_binding_failure(port) [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.463282] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] raise exception.PortBindingFailed(port_id=port['id']) [ 758.463695] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] nova.exception.PortBindingFailed: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. [ 758.463695] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] [ 758.463695] env[61911]: INFO nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Terminating instance [ 758.651850] env[61911]: DEBUG nova.network.neutron [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.715654] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Releasing lock "refresh_cache-ea27990b-d194-4b81-b18b-37804c22ceb3" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.716047] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 758.716124] env[61911]: DEBUG nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 758.716256] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.734261] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.742196] env[61911]: DEBUG nova.network.neutron [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.850658] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff19a346-fa3d-4ee6-9050-d8eee5f3ceec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.858812] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c054431-abed-4b74-860c-a68eecdb2063 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.897453] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c1b037-28a3-4023-8062-9d3270da9ab4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.905093] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d27280-eaa3-47af-bf2f-b861bf1ec739 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.918999] env[61911]: DEBUG nova.compute.provider_tree [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.967094] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.236598] env[61911]: DEBUG nova.network.neutron [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.243408] env[61911]: DEBUG oslo_concurrency.lockutils [req-db83040b-9abc-44cc-b7f2-817391ee85a3 req-b090eac1-4d85-4210-9cfa-c6b308f3ba87 service nova] Releasing lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.243831] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.244033] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.421901] env[61911]: DEBUG nova.scheduler.client.report [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 759.740549] env[61911]: INFO nova.compute.manager [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] [instance: ea27990b-d194-4b81-b18b-37804c22ceb3] Took 1.02 seconds to deallocate network for instance. [ 759.768400] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.874601] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.927373] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.929108] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 759.932179] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.895s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.187880] env[61911]: DEBUG nova.compute.manager [req-3ed569b2-19ef-407a-af8f-75057b8446b4 req-d788dcaf-9a7b-4780-aa1b-a069fb9d5cc8 service nova] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Received event network-vif-deleted-d8b0c98e-b383-4175-a292-5a81ca0174a1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 760.377064] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.377553] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 760.377776] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.378132] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec9b6ce8-e763-4ea9-a945-7d1f85bc0708 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.391993] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7efbce-6fe5-4bdf-8496-8bef21d9ff23 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.413774] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aa8d4787-0638-4bc9-99aa-01678272f59a could not be found. [ 760.414049] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.414279] env[61911]: INFO nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 760.414554] env[61911]: DEBUG oslo.service.loopingcall [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.414804] env[61911]: DEBUG nova.compute.manager [-] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 760.414928] env[61911]: DEBUG nova.network.neutron [-] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.429282] env[61911]: DEBUG nova.network.neutron [-] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.436882] env[61911]: DEBUG nova.compute.utils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.440963] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 760.442125] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.492190] env[61911]: DEBUG nova.policy [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c57bc181d3f746d39b8beb5e8aae3839', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4806147e49484ae2887767632c518e6d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.740093] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e557df-edd6-492b-9aa9-08defadb41ff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.748092] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9898d9d2-3c91-4e21-b7a9-65e0b0839422 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.781479] env[61911]: INFO nova.scheduler.client.report [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Deleted allocations for instance ea27990b-d194-4b81-b18b-37804c22ceb3 [ 760.787818] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Successfully created port: b7584005-a080-4d23-b5ec-2d0c74e824d5 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.789771] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2eeb67-a10c-47db-bb58-ae252fd43787 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.802319] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36479753-7764-439c-8cbd-c9c0379891bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.815021] env[61911]: DEBUG nova.compute.provider_tree [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.931365] env[61911]: DEBUG nova.network.neutron [-] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.942432] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 761.295968] env[61911]: DEBUG oslo_concurrency.lockutils [None req-15174eeb-64cf-4b73-8587-561f9ad445d6 tempest-ServersTestJSON-1820650719 tempest-ServersTestJSON-1820650719-project-member] Lock "ea27990b-d194-4b81-b18b-37804c22ceb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.196s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.321586] env[61911]: DEBUG nova.scheduler.client.report [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.439764] env[61911]: INFO nova.compute.manager [-] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Took 1.02 seconds to deallocate network for instance. [ 761.441130] env[61911]: DEBUG nova.compute.claims [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 761.441286] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.629858] env[61911]: ERROR nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. [ 761.629858] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.629858] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.629858] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.629858] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.629858] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.629858] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.629858] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.629858] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.629858] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 761.629858] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.629858] env[61911]: ERROR nova.compute.manager raise self.value [ 761.629858] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.629858] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.629858] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.629858] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.630675] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.630675] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.630675] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. [ 761.630675] env[61911]: ERROR nova.compute.manager [ 761.630675] env[61911]: Traceback (most recent call last): [ 761.630675] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.630675] env[61911]: listener.cb(fileno) [ 761.630675] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.630675] env[61911]: result = function(*args, **kwargs) [ 761.630675] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.630675] env[61911]: return func(*args, **kwargs) [ 761.630675] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 761.630675] env[61911]: raise e [ 761.630675] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.630675] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 761.630675] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.630675] env[61911]: created_port_ids = self._update_ports_for_instance( [ 761.630675] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.630675] env[61911]: with excutils.save_and_reraise_exception(): [ 761.630675] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.630675] env[61911]: self.force_reraise() [ 761.630675] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.630675] env[61911]: raise self.value [ 761.630675] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.630675] env[61911]: updated_port = self._update_port( [ 761.630675] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.630675] env[61911]: _ensure_no_port_binding_failure(port) [ 761.630675] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.630675] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.631631] env[61911]: nova.exception.PortBindingFailed: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. [ 761.631631] env[61911]: Removing descriptor: 16 [ 761.799638] env[61911]: DEBUG nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 761.823961] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.824714] env[61911]: ERROR nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Traceback (most recent call last): [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self.driver.spawn(context, instance, image_meta, [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] vm_ref = self.build_virtual_machine(instance, [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.824714] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] for vif in network_info: [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return self._sync_wrapper(fn, *args, **kwargs) [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self.wait() [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self[:] = self._gt.wait() [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return self._exit_event.wait() [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] result = hub.switch() [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.825126] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return self.greenlet.switch() [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] result = function(*args, **kwargs) [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] return func(*args, **kwargs) [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] raise e [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] nwinfo = self.network_api.allocate_for_instance( [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] created_port_ids = self._update_ports_for_instance( [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] with excutils.save_and_reraise_exception(): [ 761.825510] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] self.force_reraise() [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] raise self.value [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] updated_port = self._update_port( [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] _ensure_no_port_binding_failure(port) [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] raise exception.PortBindingFailed(port_id=port['id']) [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] nova.exception.PortBindingFailed: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. [ 761.825892] env[61911]: ERROR nova.compute.manager [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] [ 761.826331] env[61911]: DEBUG nova.compute.utils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.827048] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.900s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.828460] env[61911]: INFO nova.compute.claims [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.834036] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Build of instance 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9 was re-scheduled: Binding failed for port 69ab0760-59b8-45bf-9955-8905b47a9ed1, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 761.834036] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 761.834036] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquiring lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.834036] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Acquired lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.834352] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.953693] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 761.975204] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 761.975461] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.975617] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.975814] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.975944] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.976118] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 761.976345] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 761.976475] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 761.976637] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 761.977558] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 761.977558] env[61911]: DEBUG nova.virt.hardware [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 761.977833] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3485f4d5-bfd3-4797-b890-a8af88621942 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.986343] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e747fca-2344-48c8-a08b-3487010ea857 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.000251] env[61911]: ERROR nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Traceback (most recent call last): [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] yield resources [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self.driver.spawn(context, instance, image_meta, [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] vm_ref = self.build_virtual_machine(instance, [ 762.000251] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] for vif in network_info: [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] return self._sync_wrapper(fn, *args, **kwargs) [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self.wait() [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self[:] = self._gt.wait() [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] return self._exit_event.wait() [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 762.000727] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] current.throw(*self._exc) [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] result = function(*args, **kwargs) [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] return func(*args, **kwargs) [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] raise e [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] nwinfo = self.network_api.allocate_for_instance( [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] created_port_ids = self._update_ports_for_instance( [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] with excutils.save_and_reraise_exception(): [ 762.001246] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self.force_reraise() [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] raise self.value [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] updated_port = self._update_port( [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] _ensure_no_port_binding_failure(port) [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] raise exception.PortBindingFailed(port_id=port['id']) [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] nova.exception.PortBindingFailed: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. [ 762.001774] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] [ 762.001774] env[61911]: INFO nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Terminating instance [ 762.296252] env[61911]: DEBUG nova.compute.manager [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Received event network-changed-b7584005-a080-4d23-b5ec-2d0c74e824d5 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 762.296494] env[61911]: DEBUG nova.compute.manager [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Refreshing instance network info cache due to event network-changed-b7584005-a080-4d23-b5ec-2d0c74e824d5. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 762.296723] env[61911]: DEBUG oslo_concurrency.lockutils [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] Acquiring lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.296947] env[61911]: DEBUG oslo_concurrency.lockutils [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] Acquired lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.297929] env[61911]: DEBUG nova.network.neutron [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Refreshing network info cache for port b7584005-a080-4d23-b5ec-2d0c74e824d5 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.322946] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.356431] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.498125] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.505092] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.804192] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "bafd58bc-8ae7-49a4-a039-31d328699010" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.804432] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "bafd58bc-8ae7-49a4-a039-31d328699010" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.820455] env[61911]: DEBUG nova.network.neutron [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.949023] env[61911]: DEBUG nova.network.neutron [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.004017] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Releasing lock "refresh_cache-46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.004017] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 763.004017] env[61911]: DEBUG nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 763.004017] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.021895] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.205706] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8552a36-5b55-431e-8a5a-f8fa62e6ce02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.214662] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c413206c-1fd7-42ff-b107-a39453f8732b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.246276] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e455b17d-e6b0-4909-835d-8c9a077a1fb3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.253830] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29983fbf-53ef-4e4b-99f2-cfecbe1ecfc2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.268740] env[61911]: DEBUG nova.compute.provider_tree [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.280470] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "c95707ca-7e32-42c0-914a-f0b178f07997" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.280579] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "c95707ca-7e32-42c0-914a-f0b178f07997" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.454129] env[61911]: DEBUG oslo_concurrency.lockutils [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] Releasing lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.454414] env[61911]: DEBUG nova.compute.manager [req-d188b2a6-2443-4b3d-a1c8-0685f99b1f55 req-c451212f-a629-4465-b013-5ba7d2c44602 service nova] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Received event network-vif-deleted-b7584005-a080-4d23-b5ec-2d0c74e824d5 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 763.454771] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.454946] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.529077] env[61911]: DEBUG nova.network.neutron [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.771564] env[61911]: DEBUG nova.scheduler.client.report [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 763.979558] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.032183] env[61911]: INFO nova.compute.manager [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] [instance: 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9] Took 1.03 seconds to deallocate network for instance. [ 764.069158] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.277981] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.278513] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 764.281817] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.134s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.573326] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.573326] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 764.573326] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.573326] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-385315ce-8fa4-44c5-82ee-3d4a8732b59e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.582653] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6b190c-7cc0-4383-a7ab-95732e7e239b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.603076] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a77a9374-0220-43ec-bf51-34523bc8babd could not be found. [ 764.603342] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.603705] env[61911]: INFO nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 764.604008] env[61911]: DEBUG oslo.service.loopingcall [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.604250] env[61911]: DEBUG nova.compute.manager [-] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 764.604359] env[61911]: DEBUG nova.network.neutron [-] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.621101] env[61911]: DEBUG nova.network.neutron [-] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.790383] env[61911]: DEBUG nova.compute.utils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.792071] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 764.793332] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 764.835610] env[61911]: DEBUG nova.policy [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d69df1a34404f73977420529c4eb51f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acb6d2e4729b4943a1ae92ce7abcb643', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.062876] env[61911]: INFO nova.scheduler.client.report [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Deleted allocations for instance 46b80b01-88c4-4c61-9c0e-0c1c614bbdf9 [ 765.122714] env[61911]: DEBUG nova.network.neutron [-] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.134030] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Successfully created port: 0f3cd1db-0227-4831-94ba-8c8a31aa531d {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.137786] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57c8570-5e50-4689-8679-7f81b62cff30 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.146294] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cec0b4-5bf6-48e8-85fe-64c8405262af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.183015] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc83340-aa06-43b1-bd0a-50a0eed9b638 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.187334] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bddb548-d883-43ca-84d8-d6f90601b513 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.202497] env[61911]: DEBUG nova.compute.provider_tree [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.298040] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 765.572595] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f10a9482-91df-4c5a-b113-7fbf1e6ed420 tempest-SecurityGroupsTestJSON-1987299680 tempest-SecurityGroupsTestJSON-1987299680-project-member] Lock "46b80b01-88c4-4c61-9c0e-0c1c614bbdf9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.848s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.625288] env[61911]: INFO nova.compute.manager [-] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Took 1.02 seconds to deallocate network for instance. [ 765.627631] env[61911]: DEBUG nova.compute.claims [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 765.627804] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.704573] env[61911]: DEBUG nova.scheduler.client.report [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 765.813623] env[61911]: DEBUG nova.compute.manager [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Received event network-changed-0f3cd1db-0227-4831-94ba-8c8a31aa531d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 765.813862] env[61911]: DEBUG nova.compute.manager [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Refreshing instance network info cache due to event network-changed-0f3cd1db-0227-4831-94ba-8c8a31aa531d. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 765.814073] env[61911]: DEBUG oslo_concurrency.lockutils [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] Acquiring lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.814228] env[61911]: DEBUG oslo_concurrency.lockutils [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] Acquired lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.814409] env[61911]: DEBUG nova.network.neutron [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Refreshing network info cache for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.989126] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.990453] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.993946] env[61911]: ERROR nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. [ 765.993946] env[61911]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.993946] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 765.993946] env[61911]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.993946] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.993946] env[61911]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.993946] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.993946] env[61911]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.993946] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.993946] env[61911]: ERROR nova.compute.manager self.force_reraise() [ 765.993946] env[61911]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.993946] env[61911]: ERROR nova.compute.manager raise self.value [ 765.993946] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.993946] env[61911]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.993946] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.993946] env[61911]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.994462] env[61911]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.994462] env[61911]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.994462] env[61911]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. [ 765.994462] env[61911]: ERROR nova.compute.manager [ 765.994817] env[61911]: Traceback (most recent call last): [ 765.994864] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.994864] env[61911]: listener.cb(fileno) [ 765.994864] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.994864] env[61911]: result = function(*args, **kwargs) [ 765.994989] env[61911]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.994989] env[61911]: return func(*args, **kwargs) [ 765.994989] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 765.994989] env[61911]: raise e [ 765.994989] env[61911]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 765.994989] env[61911]: nwinfo = self.network_api.allocate_for_instance( [ 765.994989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.994989] env[61911]: created_port_ids = self._update_ports_for_instance( [ 765.994989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.994989] env[61911]: with excutils.save_and_reraise_exception(): [ 765.994989] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.994989] env[61911]: self.force_reraise() [ 765.994989] env[61911]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.994989] env[61911]: raise self.value [ 765.994989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.994989] env[61911]: updated_port = self._update_port( [ 765.994989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.994989] env[61911]: _ensure_no_port_binding_failure(port) [ 765.994989] env[61911]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.994989] env[61911]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.994989] env[61911]: nova.exception.PortBindingFailed: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. [ 765.994989] env[61911]: Removing descriptor: 16 [ 766.074990] env[61911]: DEBUG nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 766.209066] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.209812] env[61911]: ERROR nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Traceback (most recent call last): [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self.driver.spawn(context, instance, image_meta, [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] vm_ref = self.build_virtual_machine(instance, [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.209812] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] for vif in network_info: [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return self._sync_wrapper(fn, *args, **kwargs) [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self.wait() [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self[:] = self._gt.wait() [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return self._exit_event.wait() [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] result = hub.switch() [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 766.210211] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return self.greenlet.switch() [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] result = function(*args, **kwargs) [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] return func(*args, **kwargs) [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] raise e [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] nwinfo = self.network_api.allocate_for_instance( [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] created_port_ids = self._update_ports_for_instance( [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] with excutils.save_and_reraise_exception(): [ 766.210609] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] self.force_reraise() [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] raise self.value [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] updated_port = self._update_port( [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] _ensure_no_port_binding_failure(port) [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] raise exception.PortBindingFailed(port_id=port['id']) [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] nova.exception.PortBindingFailed: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. [ 766.210984] env[61911]: ERROR nova.compute.manager [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] [ 766.211304] env[61911]: DEBUG nova.compute.utils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.212471] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.208s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.215531] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Build of instance fe46e0e7-c1aa-4bf5-af01-cdb04031f945 was re-scheduled: Binding failed for port d04ae9d4-e5b3-48f1-ac42-8c8651b6d94e, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 766.217125] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 766.217643] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Acquiring lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.217864] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Acquired lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.218229] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.310331] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 766.337172] env[61911]: DEBUG nova.network.neutron [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.343079] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 766.343783] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.344326] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.344653] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.344814] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.345208] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 766.345450] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 766.345622] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 766.345799] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 766.345970] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 766.346503] env[61911]: DEBUG nova.virt.hardware [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 766.348045] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f04b19-4ac0-4200-a0bc-db6dce03ad69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.357778] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc7443b-18b8-4ea0-a6e7-72d38c4e0963 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.375235] env[61911]: ERROR nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Traceback (most recent call last): [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] yield resources [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self.driver.spawn(context, instance, image_meta, [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] vm_ref = self.build_virtual_machine(instance, [ 766.375235] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] for vif in network_info: [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] return self._sync_wrapper(fn, *args, **kwargs) [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self.wait() [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self[:] = self._gt.wait() [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] return self._exit_event.wait() [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.376108] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] current.throw(*self._exc) [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] result = function(*args, **kwargs) [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] return func(*args, **kwargs) [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] raise e [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] nwinfo = self.network_api.allocate_for_instance( [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] created_port_ids = self._update_ports_for_instance( [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] with excutils.save_and_reraise_exception(): [ 766.376620] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self.force_reraise() [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] raise self.value [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] updated_port = self._update_port( [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] _ensure_no_port_binding_failure(port) [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] raise exception.PortBindingFailed(port_id=port['id']) [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] nova.exception.PortBindingFailed: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. [ 766.377037] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] [ 766.377037] env[61911]: INFO nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Terminating instance [ 766.422348] env[61911]: DEBUG nova.network.neutron [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.496914] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 766.496914] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 766.598711] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.756374] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.875381] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.880508] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Acquiring lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.927146] env[61911]: DEBUG oslo_concurrency.lockutils [req-36e95a18-caf7-4994-8286-13657edc0741 req-88e6174f-7ecd-4090-a680-1827c68232c9 service nova] Releasing lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.927146] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Acquired lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.927146] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.115450] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b4ccc65-4671-43df-98b6-ba0c7a99aad0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.124669] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0dcc7e-31b0-4c9b-8ee8-de19b8d7ef59 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.163731] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed4c1f5-43b7-441b-8388-8812dd8defa0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.171471] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5dd4e3-da4d-4c92-860f-ee736bd745ec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.185498] env[61911]: DEBUG nova.compute.provider_tree [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.377319] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Releasing lock "refresh_cache-fe46e0e7-c1aa-4bf5-af01-cdb04031f945" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.377565] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 767.377747] env[61911]: DEBUG nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 767.377913] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.395457] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.446324] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.499740] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Didn't find any instances for network info cache update. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 767.499951] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.500190] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.500366] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.500520] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.500660] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.500804] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.500985] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 767.501197] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.538566] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.691950] env[61911]: DEBUG nova.scheduler.client.report [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 767.848124] env[61911]: DEBUG nova.compute.manager [req-8862fc26-5ed2-4d0f-8ad0-2aa48b19efb8 req-46b7fce6-5f20-4116-9f81-921862f26a69 service nova] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Received event network-vif-deleted-0f3cd1db-0227-4831-94ba-8c8a31aa531d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 767.898975] env[61911]: DEBUG nova.network.neutron [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.004589] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.051120] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Releasing lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.051120] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 768.051120] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.051120] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81d7052a-d65e-462a-8a3d-c41a33ff6a96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.062077] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8068866e-80b5-4200-a1de-75b23e3347f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.082812] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d4f17919-9cc9-4e19-88eb-38a978115975 could not be found. [ 768.082812] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.083099] env[61911]: INFO nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Took 0.03 seconds to destroy the instance on the hypervisor. [ 768.083408] env[61911]: DEBUG oslo.service.loopingcall [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.083680] env[61911]: DEBUG nova.compute.manager [-] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 768.083890] env[61911]: DEBUG nova.network.neutron [-] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.100215] env[61911]: DEBUG nova.network.neutron [-] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.198731] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.986s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.199398] env[61911]: ERROR nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Traceback (most recent call last): [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self.driver.spawn(context, instance, image_meta, [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] vm_ref = self.build_virtual_machine(instance, [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.199398] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] for vif in network_info: [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return self._sync_wrapper(fn, *args, **kwargs) [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self.wait() [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self[:] = self._gt.wait() [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return self._exit_event.wait() [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] result = hub.switch() [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 768.199732] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return self.greenlet.switch() [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] result = function(*args, **kwargs) [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] return func(*args, **kwargs) [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] raise e [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] nwinfo = self.network_api.allocate_for_instance( [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] created_port_ids = self._update_ports_for_instance( [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] with excutils.save_and_reraise_exception(): [ 768.200111] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] self.force_reraise() [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] raise self.value [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] updated_port = self._update_port( [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] _ensure_no_port_binding_failure(port) [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] raise exception.PortBindingFailed(port_id=port['id']) [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] nova.exception.PortBindingFailed: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. [ 768.200483] env[61911]: ERROR nova.compute.manager [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] [ 768.203487] env[61911]: DEBUG nova.compute.utils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 768.204974] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.805s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.212540] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Build of instance b6773f2b-df86-4a4c-9b20-a314e81a7aa0 was re-scheduled: Binding failed for port a9416ca8-d180-4e95-a8d7-83834df2d258, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 768.213219] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 768.213526] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Acquiring lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.213738] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Acquired lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.213956] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.246988] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "83616524-2530-4398-9019-148e0658720d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.247244] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "83616524-2530-4398-9019-148e0658720d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.401049] env[61911]: INFO nova.compute.manager [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] [instance: fe46e0e7-c1aa-4bf5-af01-cdb04031f945] Took 1.02 seconds to deallocate network for instance. [ 768.602468] env[61911]: DEBUG nova.network.neutron [-] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.736830] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.859246] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.102010] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0339158-63ce-48f1-a74f-84b0128f3e22 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.105752] env[61911]: INFO nova.compute.manager [-] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Took 1.02 seconds to deallocate network for instance. [ 769.107805] env[61911]: DEBUG nova.compute.claims [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Aborting claim: {{(pid=61911) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.108096] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.115048] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482cc5f6-c140-4cff-93bf-1a955129c33f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.144468] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274a8a70-e185-4f26-8dc7-25aedcf38878 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.151559] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94dc619d-2d8b-4ade-b964-13255bec71a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.164990] env[61911]: DEBUG nova.compute.provider_tree [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.364707] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Releasing lock "refresh_cache-b6773f2b-df86-4a4c-9b20-a314e81a7aa0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.364966] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 769.365128] env[61911]: DEBUG nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 769.365562] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.381653] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.442472] env[61911]: INFO nova.scheduler.client.report [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Deleted allocations for instance fe46e0e7-c1aa-4bf5-af01-cdb04031f945 [ 769.668602] env[61911]: DEBUG nova.scheduler.client.report [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 769.885676] env[61911]: DEBUG nova.network.neutron [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.951771] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b8ccbe50-b4c5-41dc-ab92-b85095144098 tempest-ServerAddressesTestJSON-1460665467 tempest-ServerAddressesTestJSON-1460665467-project-member] Lock "fe46e0e7-c1aa-4bf5-af01-cdb04031f945" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.051s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.176126] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.971s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.176792] env[61911]: ERROR nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Traceback (most recent call last): [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self.driver.spawn(context, instance, image_meta, [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] vm_ref = self.build_virtual_machine(instance, [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.176792] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] for vif in network_info: [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return self._sync_wrapper(fn, *args, **kwargs) [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self.wait() [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self[:] = self._gt.wait() [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return self._exit_event.wait() [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] result = hub.switch() [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 770.177175] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return self.greenlet.switch() [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] result = function(*args, **kwargs) [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] return func(*args, **kwargs) [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] raise e [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] nwinfo = self.network_api.allocate_for_instance( [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] created_port_ids = self._update_ports_for_instance( [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] with excutils.save_and_reraise_exception(): [ 770.177932] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] self.force_reraise() [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] raise self.value [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] updated_port = self._update_port( [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] _ensure_no_port_binding_failure(port) [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] raise exception.PortBindingFailed(port_id=port['id']) [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] nova.exception.PortBindingFailed: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. [ 770.178546] env[61911]: ERROR nova.compute.manager [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] [ 770.179122] env[61911]: DEBUG nova.compute.utils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 770.179122] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.073s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.181614] env[61911]: INFO nova.compute.claims [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.185074] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Build of instance b712ffa7-19d8-47d8-8d60-48ca52735776 was re-scheduled: Binding failed for port c086fba1-2024-4244-b3aa-c7e3c3b470ae, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 770.185907] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 770.185907] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Acquiring lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.186057] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Acquired lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.186098] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.388722] env[61911]: INFO nova.compute.manager [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] [instance: b6773f2b-df86-4a4c-9b20-a314e81a7aa0] Took 1.02 seconds to deallocate network for instance. [ 770.454545] env[61911]: DEBUG nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 770.712501] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.798365] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.977066] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.301495] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Releasing lock "refresh_cache-b712ffa7-19d8-47d8-8d60-48ca52735776" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.301921] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 771.301921] env[61911]: DEBUG nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 771.302094] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.328799] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.424444] env[61911]: INFO nova.scheduler.client.report [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Deleted allocations for instance b6773f2b-df86-4a4c-9b20-a314e81a7aa0 [ 771.506725] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ff4ba4-cd98-40a3-9d38-aae5b4ac43ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.519366] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602202aa-e790-40fb-ba9d-2ab4e0bb103f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.552230] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6ff432-1208-481a-89ba-6e816fded342 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.560049] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8adf35-528f-461c-ba72-b6781f3a6108 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.577871] env[61911]: DEBUG nova.compute.provider_tree [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.832411] env[61911]: DEBUG nova.network.neutron [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.940872] env[61911]: DEBUG oslo_concurrency.lockutils [None req-80ad52e7-71c3-45ac-9571-d994d3483b7d tempest-ServersV294TestFqdnHostnames-550006994 tempest-ServersV294TestFqdnHostnames-550006994-project-member] Lock "b6773f2b-df86-4a4c-9b20-a314e81a7aa0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.176s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.081795] env[61911]: DEBUG nova.scheduler.client.report [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 772.335502] env[61911]: INFO nova.compute.manager [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] [instance: b712ffa7-19d8-47d8-8d60-48ca52735776] Took 1.03 seconds to deallocate network for instance. [ 772.444117] env[61911]: DEBUG nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 772.587623] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.588184] env[61911]: DEBUG nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 772.592186] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.587s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.597718] env[61911]: INFO nova.compute.claims [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.762670] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "65d310c6-d438-4c3d-bd44-f3b51123fe93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.762909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "65d310c6-d438-4c3d-bd44-f3b51123fe93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.974313] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.102288] env[61911]: DEBUG nova.compute.utils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.108822] env[61911]: DEBUG nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 773.109014] env[61911]: DEBUG nova.network.neutron [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 773.178855] env[61911]: DEBUG nova.policy [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9aa5356986934d12aa2fee1550641bba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c42e45bd558a4beb862a0b4a2809fbe0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.373482] env[61911]: INFO nova.scheduler.client.report [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Deleted allocations for instance b712ffa7-19d8-47d8-8d60-48ca52735776 [ 773.609128] env[61911]: DEBUG nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 773.766207] env[61911]: DEBUG nova.network.neutron [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Successfully created port: 14cd0e60-af95-417b-a8b0-6f143afa5e8b {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.892238] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ff00d0a-536e-45f6-8755-f75a29557064 tempest-ServersTestManualDisk-1809199165 tempest-ServersTestManualDisk-1809199165-project-member] Lock "b712ffa7-19d8-47d8-8d60-48ca52735776" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.448s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.977295] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c852c7-b7b1-41fc-8119-c14b984eafba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.991390] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6cfc3b-5132-4ad7-9541-1ff4cdcca4b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.023574] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee1ad28-e2f3-4626-b1fd-740fc1e03e30 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.031283] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e2ce78-c0e4-408c-87f6-f0de737e306d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.046307] env[61911]: DEBUG nova.compute.provider_tree [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.398737] env[61911]: DEBUG nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 774.552028] env[61911]: DEBUG nova.scheduler.client.report [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 774.617382] env[61911]: DEBUG nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 774.651270] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 774.651885] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.652131] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 774.652539] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.652766] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 774.652983] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 774.653549] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 774.653549] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 774.653750] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 774.653985] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 774.654675] env[61911]: DEBUG nova.virt.hardware [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 774.655935] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a84480-6cc1-429d-b8b7-4c3bc71409c2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.664472] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b285a55-5069-4706-adf0-ed30438718fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.918686] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.063048] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.063048] env[61911]: DEBUG nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 775.064536] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.623s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.547312] env[61911]: DEBUG nova.compute.manager [req-cf9f9092-2c14-4ff0-bbf2-d7ffef3bddf8 req-6708006e-def5-49fd-9599-66e046fa485c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Received event network-vif-plugged-14cd0e60-af95-417b-a8b0-6f143afa5e8b {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 775.547534] env[61911]: DEBUG oslo_concurrency.lockutils [req-cf9f9092-2c14-4ff0-bbf2-d7ffef3bddf8 req-6708006e-def5-49fd-9599-66e046fa485c service nova] Acquiring lock "21723e9a-5c6d-4142-baba-25c59a5038ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.547741] env[61911]: DEBUG oslo_concurrency.lockutils [req-cf9f9092-2c14-4ff0-bbf2-d7ffef3bddf8 req-6708006e-def5-49fd-9599-66e046fa485c service nova] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.547948] env[61911]: DEBUG oslo_concurrency.lockutils [req-cf9f9092-2c14-4ff0-bbf2-d7ffef3bddf8 req-6708006e-def5-49fd-9599-66e046fa485c service nova] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.548076] env[61911]: DEBUG nova.compute.manager [req-cf9f9092-2c14-4ff0-bbf2-d7ffef3bddf8 req-6708006e-def5-49fd-9599-66e046fa485c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] No waiting events found dispatching network-vif-plugged-14cd0e60-af95-417b-a8b0-6f143afa5e8b {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 775.548271] env[61911]: WARNING nova.compute.manager [req-cf9f9092-2c14-4ff0-bbf2-d7ffef3bddf8 req-6708006e-def5-49fd-9599-66e046fa485c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Received unexpected event network-vif-plugged-14cd0e60-af95-417b-a8b0-6f143afa5e8b for instance with vm_state building and task_state spawning. [ 775.572696] env[61911]: DEBUG nova.compute.utils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.574289] env[61911]: DEBUG nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 775.574469] env[61911]: DEBUG nova.network.neutron [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 775.661131] env[61911]: DEBUG nova.network.neutron [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Successfully updated port: 14cd0e60-af95-417b-a8b0-6f143afa5e8b {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 775.661131] env[61911]: DEBUG nova.policy [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8593a447cc994e1eb0d61a548ba98ba8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e38ba711d3e3431b9d258c91ee1ce6f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.905019] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5795f41e-b885-4e5a-ba97-fe562ae8df20 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.914435] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408e13bc-752a-4865-aaf6-b9c40b21484a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.948843] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ab5279-176e-4d61-b33c-7be39d1e81b0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.956399] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c16c63-69f1-4b0c-a539-edc72cd861b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.970053] env[61911]: DEBUG nova.compute.provider_tree [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.077620] env[61911]: DEBUG nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 776.163067] env[61911]: DEBUG nova.network.neutron [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Successfully created port: 90825141-c63e-4853-8f27-6553cc0f3ecc {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.166907] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "refresh_cache-21723e9a-5c6d-4142-baba-25c59a5038ea" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.166907] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquired lock "refresh_cache-21723e9a-5c6d-4142-baba-25c59a5038ea" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.166907] env[61911]: DEBUG nova.network.neutron [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.478524] env[61911]: DEBUG nova.scheduler.client.report [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 776.706233] env[61911]: DEBUG nova.network.neutron [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.869102] env[61911]: DEBUG nova.network.neutron [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Updating instance_info_cache with network_info: [{"id": "14cd0e60-af95-417b-a8b0-6f143afa5e8b", "address": "fa:16:3e:fe:fd:33", "network": {"id": "322e763b-15d2-4945-984b-717026fc8f75", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1336185440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c42e45bd558a4beb862a0b4a2809fbe0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14cd0e60-af", "ovs_interfaceid": "14cd0e60-af95-417b-a8b0-6f143afa5e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.984009] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.984793] env[61911]: ERROR nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Traceback (most recent call last): [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self.driver.spawn(context, instance, image_meta, [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] vm_ref = self.build_virtual_machine(instance, [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.984793] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] for vif in network_info: [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return self._sync_wrapper(fn, *args, **kwargs) [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self.wait() [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self[:] = self._gt.wait() [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return self._exit_event.wait() [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] result = hub.switch() [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 776.985224] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return self.greenlet.switch() [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] result = function(*args, **kwargs) [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] return func(*args, **kwargs) [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] raise e [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] nwinfo = self.network_api.allocate_for_instance( [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] created_port_ids = self._update_ports_for_instance( [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] with excutils.save_and_reraise_exception(): [ 776.985668] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] self.force_reraise() [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] raise self.value [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] updated_port = self._update_port( [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] _ensure_no_port_binding_failure(port) [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] raise exception.PortBindingFailed(port_id=port['id']) [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] nova.exception.PortBindingFailed: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. [ 776.986173] env[61911]: ERROR nova.compute.manager [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] [ 776.986534] env[61911]: DEBUG nova.compute.utils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.987122] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.665s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.988529] env[61911]: INFO nova.compute.claims [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.992926] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Build of instance aa8d4787-0638-4bc9-99aa-01678272f59a was re-scheduled: Binding failed for port d8b0c98e-b383-4175-a292-5a81ca0174a1, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 776.993375] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 776.993655] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.993811] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.993977] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.087904] env[61911]: DEBUG nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 777.111431] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 777.111728] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.111901] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 777.112112] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.112262] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 777.112432] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 777.112659] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 777.112820] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 777.112988] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 777.113172] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 777.113427] env[61911]: DEBUG nova.virt.hardware [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 777.114288] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515890c7-0761-4690-9264-c7556454bf19 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.122472] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19bf387-45da-4fd2-ae5f-71612d78b604 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.372604] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Releasing lock "refresh_cache-21723e9a-5c6d-4142-baba-25c59a5038ea" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.372944] env[61911]: DEBUG nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Instance network_info: |[{"id": "14cd0e60-af95-417b-a8b0-6f143afa5e8b", "address": "fa:16:3e:fe:fd:33", "network": {"id": "322e763b-15d2-4945-984b-717026fc8f75", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1336185440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c42e45bd558a4beb862a0b4a2809fbe0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14cd0e60-af", "ovs_interfaceid": "14cd0e60-af95-417b-a8b0-6f143afa5e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 777.373519] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:fd:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39a4aca0-934b-4a91-8779-6a4360c3f967', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14cd0e60-af95-417b-a8b0-6f143afa5e8b', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.381161] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Creating folder: Project (c42e45bd558a4beb862a0b4a2809fbe0). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.381412] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4cb7f450-4cb8-466e-a458-b6b8bf9d3bd1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.392478] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Created folder: Project (c42e45bd558a4beb862a0b4a2809fbe0) in parent group-v269521. [ 777.392666] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Creating folder: Instances. Parent ref: group-v269542. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.392902] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d53db139-a9c4-48b7-8fb1-54db2994b37f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.401810] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Created folder: Instances in parent group-v269542. [ 777.402036] env[61911]: DEBUG oslo.service.loopingcall [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.402217] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.402397] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1963ea4-3264-4dea-a567-2e1401a273a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.421061] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.421061] env[61911]: value = "task-1250904" [ 777.421061] env[61911]: _type = "Task" [ 777.421061] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.428017] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250904, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.513713] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.573293] env[61911]: DEBUG nova.compute.manager [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Received event network-changed-14cd0e60-af95-417b-a8b0-6f143afa5e8b {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 777.573648] env[61911]: DEBUG nova.compute.manager [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Refreshing instance network info cache due to event network-changed-14cd0e60-af95-417b-a8b0-6f143afa5e8b. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 777.574456] env[61911]: DEBUG oslo_concurrency.lockutils [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] Acquiring lock "refresh_cache-21723e9a-5c6d-4142-baba-25c59a5038ea" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.574503] env[61911]: DEBUG oslo_concurrency.lockutils [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] Acquired lock "refresh_cache-21723e9a-5c6d-4142-baba-25c59a5038ea" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.574645] env[61911]: DEBUG nova.network.neutron [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Refreshing network info cache for port 14cd0e60-af95-417b-a8b0-6f143afa5e8b {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 777.595420] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.697885] env[61911]: DEBUG nova.compute.manager [req-ce0fdc0c-765b-4860-abea-88ea84047908 req-b0110d89-7810-4a69-b17c-f4e26964da63 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Received event network-vif-plugged-90825141-c63e-4853-8f27-6553cc0f3ecc {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 777.697976] env[61911]: DEBUG oslo_concurrency.lockutils [req-ce0fdc0c-765b-4860-abea-88ea84047908 req-b0110d89-7810-4a69-b17c-f4e26964da63 service nova] Acquiring lock "66472f43-537d-4eb3-8d49-d40627a8809d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.698199] env[61911]: DEBUG oslo_concurrency.lockutils [req-ce0fdc0c-765b-4860-abea-88ea84047908 req-b0110d89-7810-4a69-b17c-f4e26964da63 service nova] Lock "66472f43-537d-4eb3-8d49-d40627a8809d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.698370] env[61911]: DEBUG oslo_concurrency.lockutils [req-ce0fdc0c-765b-4860-abea-88ea84047908 req-b0110d89-7810-4a69-b17c-f4e26964da63 service nova] Lock "66472f43-537d-4eb3-8d49-d40627a8809d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.698535] env[61911]: DEBUG nova.compute.manager [req-ce0fdc0c-765b-4860-abea-88ea84047908 req-b0110d89-7810-4a69-b17c-f4e26964da63 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] No waiting events found dispatching network-vif-plugged-90825141-c63e-4853-8f27-6553cc0f3ecc {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 777.698699] env[61911]: WARNING nova.compute.manager [req-ce0fdc0c-765b-4860-abea-88ea84047908 req-b0110d89-7810-4a69-b17c-f4e26964da63 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Received unexpected event network-vif-plugged-90825141-c63e-4853-8f27-6553cc0f3ecc for instance with vm_state building and task_state spawning. [ 777.834113] env[61911]: DEBUG nova.network.neutron [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Successfully updated port: 90825141-c63e-4853-8f27-6553cc0f3ecc {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.930707] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250904, 'name': CreateVM_Task, 'duration_secs': 0.312751} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.930833] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 777.937661] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.937839] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.938177] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.939024] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17b6474a-1415-42bc-8d19-e813aa181955 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.942994] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 777.942994] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5233f1f5-1986-32bd-f6d8-de9059ee47e8" [ 777.942994] env[61911]: _type = "Task" [ 777.942994] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.951738] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5233f1f5-1986-32bd-f6d8-de9059ee47e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.098839] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-aa8d4787-0638-4bc9-99aa-01678272f59a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.099096] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 778.099281] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 778.099487] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.114027] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.222874] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8fbe39-e3c2-4d02-a186-6ee14623c795 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.230577] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8223e59b-1608-42a6-b6ab-a1db2dddb179 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.262856] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c436188-317e-4543-b677-cc7a831ed612 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.270175] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6db9e0-e3c6-40d6-a23c-c0a1a6c1f0b5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.283065] env[61911]: DEBUG nova.compute.provider_tree [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.336890] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.337140] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.339845] env[61911]: DEBUG nova.network.neutron [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.408767] env[61911]: DEBUG nova.network.neutron [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Updated VIF entry in instance network info cache for port 14cd0e60-af95-417b-a8b0-6f143afa5e8b. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 778.409534] env[61911]: DEBUG nova.network.neutron [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Updating instance_info_cache with network_info: [{"id": "14cd0e60-af95-417b-a8b0-6f143afa5e8b", "address": "fa:16:3e:fe:fd:33", "network": {"id": "322e763b-15d2-4945-984b-717026fc8f75", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1336185440-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c42e45bd558a4beb862a0b4a2809fbe0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14cd0e60-af", "ovs_interfaceid": "14cd0e60-af95-417b-a8b0-6f143afa5e8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.452988] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5233f1f5-1986-32bd-f6d8-de9059ee47e8, 'name': SearchDatastore_Task, 'duration_secs': 0.012104} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.453320] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.453563] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.453796] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.453959] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.454123] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.454650] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5167fdb3-d1c3-4c2a-aaac-3dea3ccb66a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.463025] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.463025] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 778.463604] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25a8715d-87d2-4507-9691-2f74a47176cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.470280] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 778.470280] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52603df8-92d3-af6e-59cd-78b9139c0db4" [ 778.470280] env[61911]: _type = "Task" [ 778.470280] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.477863] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52603df8-92d3-af6e-59cd-78b9139c0db4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.616547] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.786399] env[61911]: DEBUG nova.scheduler.client.report [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 778.884398] env[61911]: DEBUG nova.network.neutron [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.914059] env[61911]: DEBUG oslo_concurrency.lockutils [req-ee074993-686a-4bd7-a9d5-2e298ce31ac2 req-f0d1c10f-fb5c-43e2-8d5b-6fafbe4c3f8c service nova] Releasing lock "refresh_cache-21723e9a-5c6d-4142-baba-25c59a5038ea" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.981334] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52603df8-92d3-af6e-59cd-78b9139c0db4, 'name': SearchDatastore_Task, 'duration_secs': 0.008797} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.982196] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3c1c1f8-618f-436a-a0e8-816add4527b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.987748] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 778.987748] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad797d-9f0f-68db-9d44-f12dfb2b025e" [ 778.987748] env[61911]: _type = "Task" [ 778.987748] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.996073] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad797d-9f0f-68db-9d44-f12dfb2b025e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.119520] env[61911]: INFO nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: aa8d4787-0638-4bc9-99aa-01678272f59a] Took 1.02 seconds to deallocate network for instance. [ 779.205026] env[61911]: DEBUG nova.network.neutron [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Updating instance_info_cache with network_info: [{"id": "90825141-c63e-4853-8f27-6553cc0f3ecc", "address": "fa:16:3e:94:fb:80", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90825141-c6", "ovs_interfaceid": "90825141-c63e-4853-8f27-6553cc0f3ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.291071] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.292018] env[61911]: DEBUG nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 779.294120] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.666s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.498724] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad797d-9f0f-68db-9d44-f12dfb2b025e, 'name': SearchDatastore_Task, 'duration_secs': 0.009514} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.499062] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.499350] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 21723e9a-5c6d-4142-baba-25c59a5038ea/21723e9a-5c6d-4142-baba-25c59a5038ea.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.499606] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e952ab13-b5f2-42cc-a132-66f8ef8134a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.505895] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 779.505895] env[61911]: value = "task-1250905" [ 779.505895] env[61911]: _type = "Task" [ 779.505895] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.513097] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.708123] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.708424] env[61911]: DEBUG nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance network_info: |[{"id": "90825141-c63e-4853-8f27-6553cc0f3ecc", "address": "fa:16:3e:94:fb:80", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90825141-c6", "ovs_interfaceid": "90825141-c63e-4853-8f27-6553cc0f3ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 779.708853] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:fb:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90825141-c63e-4853-8f27-6553cc0f3ecc', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.716676] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating folder: Project (e38ba711d3e3431b9d258c91ee1ce6f1). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.717836] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55a9fc27-2530-43d6-8207-c9ca002787f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.725807] env[61911]: DEBUG nova.compute.manager [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Received event network-changed-90825141-c63e-4853-8f27-6553cc0f3ecc {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 779.726014] env[61911]: DEBUG nova.compute.manager [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Refreshing instance network info cache due to event network-changed-90825141-c63e-4853-8f27-6553cc0f3ecc. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 779.726226] env[61911]: DEBUG oslo_concurrency.lockutils [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] Acquiring lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.726336] env[61911]: DEBUG oslo_concurrency.lockutils [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] Acquired lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.726493] env[61911]: DEBUG nova.network.neutron [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Refreshing network info cache for port 90825141-c63e-4853-8f27-6553cc0f3ecc {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.729375] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created folder: Project (e38ba711d3e3431b9d258c91ee1ce6f1) in parent group-v269521. [ 779.729493] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating folder: Instances. Parent ref: group-v269545. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.729866] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f63c8f0-2d19-44f3-b2a4-519fa78d77ff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.739146] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created folder: Instances in parent group-v269545. [ 779.739422] env[61911]: DEBUG oslo.service.loopingcall [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.739589] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 779.739798] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf05ad1f-b21b-449b-bf37-cdce8c0e84ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.760707] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.760707] env[61911]: value = "task-1250908" [ 779.760707] env[61911]: _type = "Task" [ 779.760707] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.769098] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250908, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.798729] env[61911]: DEBUG nova.compute.utils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.805478] env[61911]: DEBUG nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 779.805673] env[61911]: DEBUG nova.network.neutron [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.892643] env[61911]: DEBUG nova.policy [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8593a447cc994e1eb0d61a548ba98ba8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e38ba711d3e3431b9d258c91ee1ce6f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.017209] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476963} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.019970] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 21723e9a-5c6d-4142-baba-25c59a5038ea/21723e9a-5c6d-4142-baba-25c59a5038ea.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.020429] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.020586] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38bf861a-f406-447d-bc8f-547275e6cd7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.026674] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 780.026674] env[61911]: value = "task-1250909" [ 780.026674] env[61911]: _type = "Task" [ 780.026674] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.037309] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.082505] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a8b384-36ac-4ccd-9161-a13fe73e2a38 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.089575] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e22285-bef6-4466-b4d3-f2f261631ebf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.118909] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a992a4-c5f9-4aee-a819-02cca6b9bff5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.126352] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39197462-d5e0-4715-bd02-da1aa30ca38c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.142672] env[61911]: DEBUG nova.compute.provider_tree [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.153117] env[61911]: INFO nova.scheduler.client.report [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Deleted allocations for instance aa8d4787-0638-4bc9-99aa-01678272f59a [ 780.234312] env[61911]: DEBUG nova.network.neutron [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Successfully created port: 2e5e758d-6c28-4301-b6ec-99bd3a385c91 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.274931] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250908, 'name': CreateVM_Task, 'duration_secs': 0.374617} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.274931] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 780.277529] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.277529] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.277529] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 780.277529] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0eec77c-68b2-4705-bea0-67c03dceabca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.281028] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 780.281028] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c8fe08-3d99-26e1-4a48-1d5f15ad324b" [ 780.281028] env[61911]: _type = "Task" [ 780.281028] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.288353] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c8fe08-3d99-26e1-4a48-1d5f15ad324b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.305713] env[61911]: DEBUG nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 780.447351] env[61911]: DEBUG nova.network.neutron [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Updated VIF entry in instance network info cache for port 90825141-c63e-4853-8f27-6553cc0f3ecc. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 780.447720] env[61911]: DEBUG nova.network.neutron [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Updating instance_info_cache with network_info: [{"id": "90825141-c63e-4853-8f27-6553cc0f3ecc", "address": "fa:16:3e:94:fb:80", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90825141-c6", "ovs_interfaceid": "90825141-c63e-4853-8f27-6553cc0f3ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.536088] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062801} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.536381] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.537170] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c8f5b0-8a8c-496b-87b0-e80ce86680dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.558649] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 21723e9a-5c6d-4142-baba-25c59a5038ea/21723e9a-5c6d-4142-baba-25c59a5038ea.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.558909] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f516bcd-dbdd-45bf-a3c1-a7bde16dd2e0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.579843] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 780.579843] env[61911]: value = "task-1250910" [ 780.579843] env[61911]: _type = "Task" [ 780.579843] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.588145] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250910, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.648483] env[61911]: DEBUG nova.scheduler.client.report [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 780.663251] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "aa8d4787-0638-4bc9-99aa-01678272f59a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.472s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.790841] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c8fe08-3d99-26e1-4a48-1d5f15ad324b, 'name': SearchDatastore_Task, 'duration_secs': 0.0089} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.791792] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.791792] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.791997] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.791997] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.792144] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.792390] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3df44dc4-cbb3-4217-a50a-74154911f3da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.799623] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.799797] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 780.800469] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-089388e5-bfe3-4d24-a7ad-581cccb3d433 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.804924] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 780.804924] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b1a01c-0092-c5f6-47fa-6821f50a9750" [ 780.804924] env[61911]: _type = "Task" [ 780.804924] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.815120] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b1a01c-0092-c5f6-47fa-6821f50a9750, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.951074] env[61911]: DEBUG oslo_concurrency.lockutils [req-22caf585-3f56-49a8-9c80-0a01f0d0ecc0 req-5fe46cd2-02ae-4162-956e-3e9528ad0ca9 service nova] Releasing lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.090056] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250910, 'name': ReconfigVM_Task, 'duration_secs': 0.310856} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.090525] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 21723e9a-5c6d-4142-baba-25c59a5038ea/21723e9a-5c6d-4142-baba-25c59a5038ea.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.091007] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e48f1229-78e9-4f44-96e8-ebf975cbe9bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.096799] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 781.096799] env[61911]: value = "task-1250911" [ 781.096799] env[61911]: _type = "Task" [ 781.096799] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.104468] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250911, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.154710] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.860s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.155531] env[61911]: ERROR nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Traceback (most recent call last): [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self.driver.spawn(context, instance, image_meta, [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] vm_ref = self.build_virtual_machine(instance, [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.155531] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] for vif in network_info: [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] return self._sync_wrapper(fn, *args, **kwargs) [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self.wait() [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self[:] = self._gt.wait() [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] return self._exit_event.wait() [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] current.throw(*self._exc) [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.156026] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] result = function(*args, **kwargs) [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] return func(*args, **kwargs) [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] raise e [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] nwinfo = self.network_api.allocate_for_instance( [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] created_port_ids = self._update_ports_for_instance( [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] with excutils.save_and_reraise_exception(): [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] self.force_reraise() [ 781.156467] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] raise self.value [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] updated_port = self._update_port( [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] _ensure_no_port_binding_failure(port) [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] raise exception.PortBindingFailed(port_id=port['id']) [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] nova.exception.PortBindingFailed: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. [ 781.156911] env[61911]: ERROR nova.compute.manager [instance: a77a9374-0220-43ec-bf51-34523bc8babd] [ 781.156911] env[61911]: DEBUG nova.compute.utils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.157701] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.559s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.159180] env[61911]: INFO nova.compute.claims [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.161844] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Build of instance a77a9374-0220-43ec-bf51-34523bc8babd was re-scheduled: Binding failed for port b7584005-a080-4d23-b5ec-2d0c74e824d5, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 781.162297] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 781.162518] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquiring lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.162669] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Acquired lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.162844] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.166746] env[61911]: DEBUG nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 781.317815] env[61911]: DEBUG nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 781.319890] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b1a01c-0092-c5f6-47fa-6821f50a9750, 'name': SearchDatastore_Task, 'duration_secs': 0.009163} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.320733] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba928b25-aa9d-42e5-b71a-71b5a72cfa56 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.325666] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 781.325666] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528b0fe0-0e74-8952-c6f1-7ce93e4ce24c" [ 781.325666] env[61911]: _type = "Task" [ 781.325666] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.333620] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528b0fe0-0e74-8952-c6f1-7ce93e4ce24c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.341770] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 781.342015] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.342181] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 781.342373] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.342580] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 781.342780] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 781.343059] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 781.343233] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 781.343469] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 781.343665] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 781.343859] env[61911]: DEBUG nova.virt.hardware [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 781.344637] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6bd708-dac4-4c2e-9579-71df08355157 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.351930] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6751fe-cb8b-4def-849b-aa6e0b03ac82 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.606423] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250911, 'name': Rename_Task, 'duration_secs': 0.135423} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.606700] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 781.606980] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8678b505-80d2-470b-a1ee-bba4115e1c00 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.612477] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 781.612477] env[61911]: value = "task-1250912" [ 781.612477] env[61911]: _type = "Task" [ 781.612477] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.619570] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250912, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.686947] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.689330] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.772470] env[61911]: DEBUG nova.compute.manager [req-b9847d63-0b08-45fe-b1a0-bc711cb3516c req-8f2c7465-10d1-4a7d-99f5-19af3265dbda service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Received event network-vif-plugged-2e5e758d-6c28-4301-b6ec-99bd3a385c91 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 781.773055] env[61911]: DEBUG oslo_concurrency.lockutils [req-b9847d63-0b08-45fe-b1a0-bc711cb3516c req-8f2c7465-10d1-4a7d-99f5-19af3265dbda service nova] Acquiring lock "267075d7-c3d8-4520-b7d9-67504097f823-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.773055] env[61911]: DEBUG oslo_concurrency.lockutils [req-b9847d63-0b08-45fe-b1a0-bc711cb3516c req-8f2c7465-10d1-4a7d-99f5-19af3265dbda service nova] Lock "267075d7-c3d8-4520-b7d9-67504097f823-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.773225] env[61911]: DEBUG oslo_concurrency.lockutils [req-b9847d63-0b08-45fe-b1a0-bc711cb3516c req-8f2c7465-10d1-4a7d-99f5-19af3265dbda service nova] Lock "267075d7-c3d8-4520-b7d9-67504097f823-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.773599] env[61911]: DEBUG nova.compute.manager [req-b9847d63-0b08-45fe-b1a0-bc711cb3516c req-8f2c7465-10d1-4a7d-99f5-19af3265dbda service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] No waiting events found dispatching network-vif-plugged-2e5e758d-6c28-4301-b6ec-99bd3a385c91 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 781.773599] env[61911]: WARNING nova.compute.manager [req-b9847d63-0b08-45fe-b1a0-bc711cb3516c req-8f2c7465-10d1-4a7d-99f5-19af3265dbda service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Received unexpected event network-vif-plugged-2e5e758d-6c28-4301-b6ec-99bd3a385c91 for instance with vm_state building and task_state spawning. [ 781.774745] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.836148] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528b0fe0-0e74-8952-c6f1-7ce93e4ce24c, 'name': SearchDatastore_Task, 'duration_secs': 0.009229} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.836574] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.836944] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 781.837337] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40437578-dfe3-41d9-8856-588d54d62b6e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.847147] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 781.847147] env[61911]: value = "task-1250913" [ 781.847147] env[61911]: _type = "Task" [ 781.847147] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.853732] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250913, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.880883] env[61911]: DEBUG nova.network.neutron [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Successfully updated port: 2e5e758d-6c28-4301-b6ec-99bd3a385c91 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.127528] env[61911]: DEBUG oslo_vmware.api [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250912, 'name': PowerOnVM_Task, 'duration_secs': 0.425933} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.127982] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 782.128307] env[61911]: INFO nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Took 7.51 seconds to spawn the instance on the hypervisor. [ 782.128633] env[61911]: DEBUG nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 782.129856] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138de2ff-3add-4dc3-8a65-b66d17e28c24 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.277184] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Releasing lock "refresh_cache-a77a9374-0220-43ec-bf51-34523bc8babd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.277474] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 782.277674] env[61911]: DEBUG nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 782.277883] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.296151] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.355393] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250913, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503208} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.357844] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 782.358081] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.358484] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fcf1b3f-3d5b-44c0-94a4-1875ad845594 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.364684] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 782.364684] env[61911]: value = "task-1250914" [ 782.364684] env[61911]: _type = "Task" [ 782.364684] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.374564] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250914, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.383533] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "refresh_cache-267075d7-c3d8-4520-b7d9-67504097f823" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.383688] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "refresh_cache-267075d7-c3d8-4520-b7d9-67504097f823" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.383841] env[61911]: DEBUG nova.network.neutron [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.435740] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef38c564-4c00-4d09-84b2-613e50a4c7ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.443016] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598d8890-1b78-43ca-a007-137fd64c0927 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.472471] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bbc958-cfe6-4661-a20b-e29e1c0d0cdc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.479681] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d1d166-6799-4b4d-b5ad-34dd6ab4a6ec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.492645] env[61911]: DEBUG nova.compute.provider_tree [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.655559] env[61911]: INFO nova.compute.manager [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Took 26.57 seconds to build instance. [ 782.798825] env[61911]: DEBUG nova.network.neutron [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.876179] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250914, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.321386} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.876438] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 782.878464] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13c9c49-78b8-4a89-a031-2c995ad65883 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.899552] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 782.900042] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9122d239-88c4-4a2c-af60-4801191fe70b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.918631] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 782.918631] env[61911]: value = "task-1250915" [ 782.918631] env[61911]: _type = "Task" [ 782.918631] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.927016] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250915, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.960524] env[61911]: DEBUG nova.network.neutron [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.996156] env[61911]: DEBUG nova.scheduler.client.report [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 783.137049] env[61911]: DEBUG nova.network.neutron [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Updating instance_info_cache with network_info: [{"id": "2e5e758d-6c28-4301-b6ec-99bd3a385c91", "address": "fa:16:3e:95:f0:e1", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e5e758d-6c", "ovs_interfaceid": "2e5e758d-6c28-4301-b6ec-99bd3a385c91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.160173] env[61911]: DEBUG oslo_concurrency.lockutils [None req-26ad841f-b4ff-4239-8e22-751d2918e7bb tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.298s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.301984] env[61911]: INFO nova.compute.manager [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] [instance: a77a9374-0220-43ec-bf51-34523bc8babd] Took 1.02 seconds to deallocate network for instance. [ 783.429109] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.501459] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.344s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.502034] env[61911]: DEBUG nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 783.504651] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.500s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.504827] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.504979] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 783.505273] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.397s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.508609] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4623ca-9b33-4313-b92b-37fc86b13843 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.516184] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1128813b-3ee2-40e2-a280-81152e37f460 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.529395] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9b6d3e-069e-48aa-a660-2ccc8e88c448 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.535807] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e3d253-848b-4ce4-bc20-428ec4b03d69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.269208] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "refresh_cache-267075d7-c3d8-4520-b7d9-67504097f823" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.269521] env[61911]: DEBUG nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Instance network_info: |[{"id": "2e5e758d-6c28-4301-b6ec-99bd3a385c91", "address": "fa:16:3e:95:f0:e1", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e5e758d-6c", "ovs_interfaceid": "2e5e758d-6c28-4301-b6ec-99bd3a385c91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 784.269860] env[61911]: DEBUG nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 784.272267] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "21723e9a-5c6d-4142-baba-25c59a5038ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.272482] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.272676] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "21723e9a-5c6d-4142-baba-25c59a5038ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.272851] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.273014] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.278736] env[61911]: DEBUG nova.compute.utils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.282177] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181539MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 784.282314] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.284199] env[61911]: INFO nova.compute.manager [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Terminating instance [ 784.284599] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:f0:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e5e758d-6c28-4301-b6ec-99bd3a385c91', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 784.291725] env[61911]: DEBUG oslo.service.loopingcall [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.292950] env[61911]: DEBUG nova.compute.manager [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Received event network-changed-2e5e758d-6c28-4301-b6ec-99bd3a385c91 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 784.293141] env[61911]: DEBUG nova.compute.manager [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Refreshing instance network info cache due to event network-changed-2e5e758d-6c28-4301-b6ec-99bd3a385c91. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 784.293339] env[61911]: DEBUG oslo_concurrency.lockutils [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] Acquiring lock "refresh_cache-267075d7-c3d8-4520-b7d9-67504097f823" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.293546] env[61911]: DEBUG oslo_concurrency.lockutils [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] Acquired lock "refresh_cache-267075d7-c3d8-4520-b7d9-67504097f823" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.293666] env[61911]: DEBUG nova.network.neutron [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Refreshing network info cache for port 2e5e758d-6c28-4301-b6ec-99bd3a385c91 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 784.298543] env[61911]: DEBUG nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 784.298543] env[61911]: DEBUG nova.network.neutron [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 784.300694] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 784.301751] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06457a6d-e0d0-4c0f-b5f4-452b71762bb2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.323938] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250915, 'name': ReconfigVM_Task, 'duration_secs': 1.127436} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.323938] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.324409] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea49c9cf-b0bf-4677-9044-b665be465836 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.328306] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 784.328306] env[61911]: value = "task-1250916" [ 784.328306] env[61911]: _type = "Task" [ 784.328306] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.333033] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 784.333033] env[61911]: value = "task-1250917" [ 784.333033] env[61911]: _type = "Task" [ 784.333033] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.340670] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250916, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.346153] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250917, 'name': Rename_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.378367] env[61911]: DEBUG nova.policy [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de88f0838e5f471b9c67f8d85d21a87f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3fa4a98db1e4f39831a88402ec1ef6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.567156] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ece628-78f1-4d1a-99b5-671a6894e054 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.575580] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08967516-b7cc-4377-bef7-c9ae2be111db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.608924] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5598bf4-4319-43a0-b92c-6e26080fe220 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.616783] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c06d957-3c39-4d1c-afc8-a92d69b7b531 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.630691] env[61911]: DEBUG nova.compute.provider_tree [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.786165] env[61911]: DEBUG nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 784.804022] env[61911]: DEBUG nova.compute.manager [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 784.804312] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.805199] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98afd9c6-2f28-4969-a6ee-4a801cd813ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.810981] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.812821] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 784.813083] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a00f0e7f-145e-464d-9e1e-0869fd326821 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.818868] env[61911]: DEBUG oslo_vmware.api [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 784.818868] env[61911]: value = "task-1250918" [ 784.818868] env[61911]: _type = "Task" [ 784.818868] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.819755] env[61911]: INFO nova.scheduler.client.report [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Deleted allocations for instance a77a9374-0220-43ec-bf51-34523bc8babd [ 784.844322] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250916, 'name': CreateVM_Task, 'duration_secs': 0.372798} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.844322] env[61911]: DEBUG oslo_vmware.api [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250918, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.846796] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 784.847126] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.847293] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.847617] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 784.850901] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f4bbd3a-c444-43f0-8109-16dff7ad2cca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.853089] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250917, 'name': Rename_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.855630] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 784.855630] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b88ab6-3b6f-f62e-fcea-5178cecc90c1" [ 784.855630] env[61911]: _type = "Task" [ 784.855630] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.866208] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b88ab6-3b6f-f62e-fcea-5178cecc90c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.917852] env[61911]: DEBUG nova.network.neutron [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Successfully created port: 6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.027351] env[61911]: DEBUG nova.network.neutron [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Updated VIF entry in instance network info cache for port 2e5e758d-6c28-4301-b6ec-99bd3a385c91. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 785.027351] env[61911]: DEBUG nova.network.neutron [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Updating instance_info_cache with network_info: [{"id": "2e5e758d-6c28-4301-b6ec-99bd3a385c91", "address": "fa:16:3e:95:f0:e1", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e5e758d-6c", "ovs_interfaceid": "2e5e758d-6c28-4301-b6ec-99bd3a385c91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.135521] env[61911]: DEBUG nova.scheduler.client.report [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 785.330361] env[61911]: DEBUG oslo_concurrency.lockutils [None req-946d72b4-f810-4e53-8639-3ce0e41de405 tempest-MultipleCreateTestJSON-1451432078 tempest-MultipleCreateTestJSON-1451432078-project-member] Lock "a77a9374-0220-43ec-bf51-34523bc8babd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.106s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.335695] env[61911]: DEBUG oslo_vmware.api [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250918, 'name': PowerOffVM_Task, 'duration_secs': 0.341267} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.341287] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 785.341509] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 785.342319] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8c11f15-9b20-4ae7-83dd-5bf5302b15c8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.348619] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250917, 'name': Rename_Task, 'duration_secs': 0.795441} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.348869] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 785.349084] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2ba6af8-f301-4cc4-a663-65050f4481d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.355191] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 785.355191] env[61911]: value = "task-1250920" [ 785.355191] env[61911]: _type = "Task" [ 785.355191] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.364833] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.367967] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b88ab6-3b6f-f62e-fcea-5178cecc90c1, 'name': SearchDatastore_Task, 'duration_secs': 0.010139} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.368242] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.368463] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 785.368687] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.368833] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.369017] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 785.369242] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8710627d-d678-4238-9ade-e9be97635fe6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.376473] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 785.376647] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 785.377330] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-768cc26b-6299-4839-97dc-080dfa126a7d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.382245] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 785.382245] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b00a56-b389-d3f2-e7ae-89554d2d1269" [ 785.382245] env[61911]: _type = "Task" [ 785.382245] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.389475] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b00a56-b389-d3f2-e7ae-89554d2d1269, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.529392] env[61911]: DEBUG oslo_concurrency.lockutils [req-7f46980e-1ef7-4cec-aa39-eeef1b3b4728 req-9e4089e8-d502-4aa7-a732-18986e08ff50 service nova] Releasing lock "refresh_cache-267075d7-c3d8-4520-b7d9-67504097f823" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.639424] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.134s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.639910] env[61911]: ERROR nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Traceback (most recent call last): [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self.driver.spawn(context, instance, image_meta, [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self._vmops.spawn(context, instance, image_meta, injected_files, [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] vm_ref = self.build_virtual_machine(instance, [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] vif_infos = vmwarevif.get_vif_info(self._session, [ 785.639910] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] for vif in network_info: [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] return self._sync_wrapper(fn, *args, **kwargs) [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self.wait() [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self[:] = self._gt.wait() [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] return self._exit_event.wait() [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] current.throw(*self._exc) [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 785.640357] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] result = function(*args, **kwargs) [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] return func(*args, **kwargs) [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] raise e [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] nwinfo = self.network_api.allocate_for_instance( [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] created_port_ids = self._update_ports_for_instance( [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] with excutils.save_and_reraise_exception(): [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] self.force_reraise() [ 785.641407] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] raise self.value [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] updated_port = self._update_port( [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] _ensure_no_port_binding_failure(port) [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] raise exception.PortBindingFailed(port_id=port['id']) [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] nova.exception.PortBindingFailed: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. [ 785.641974] env[61911]: ERROR nova.compute.manager [instance: d4f17919-9cc9-4e19-88eb-38a978115975] [ 785.641974] env[61911]: DEBUG nova.compute.utils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. {{(pid=61911) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 785.642519] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.665s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.643411] env[61911]: INFO nova.compute.claims [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.646453] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Build of instance d4f17919-9cc9-4e19-88eb-38a978115975 was re-scheduled: Binding failed for port 0f3cd1db-0227-4831-94ba-8c8a31aa531d, please check neutron logs for more information. {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 785.646565] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Unplugging VIFs for instance {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 785.646736] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Acquiring lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.646885] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Acquired lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.647054] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.769055] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 785.769299] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 785.769492] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Deleting the datastore file [datastore1] 21723e9a-5c6d-4142-baba-25c59a5038ea {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.769791] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f27d272-ea9d-4e5f-9700-792e069c237c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.777334] env[61911]: DEBUG oslo_vmware.api [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for the task: (returnval){ [ 785.777334] env[61911]: value = "task-1250921" [ 785.777334] env[61911]: _type = "Task" [ 785.777334] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.786356] env[61911]: DEBUG oslo_vmware.api [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.799765] env[61911]: DEBUG nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 785.828428] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 785.828742] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.828945] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 785.829164] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.829353] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 785.829528] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 785.829780] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 785.829968] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 785.830256] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 785.830391] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 785.830605] env[61911]: DEBUG nova.virt.hardware [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 785.831583] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db98239-17c1-4730-8db6-12e28aab591c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.834671] env[61911]: DEBUG nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 785.843260] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27ed291-7ead-427d-96f1-cf70baad0dbe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.864560] env[61911]: DEBUG oslo_vmware.api [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250920, 'name': PowerOnVM_Task, 'duration_secs': 0.499121} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.864786] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 785.865029] env[61911]: INFO nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Took 8.78 seconds to spawn the instance on the hypervisor. [ 785.865188] env[61911]: DEBUG nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 785.865950] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e66e60d-5cb0-4b41-b605-b0793669729c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.892850] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b00a56-b389-d3f2-e7ae-89554d2d1269, 'name': SearchDatastore_Task, 'duration_secs': 0.008544} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.893658] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e3e9830-b187-42f1-9214-302356e83544 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.899318] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 785.899318] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52226bc7-ec48-bdf2-0fcc-106e4df64f0f" [ 785.899318] env[61911]: _type = "Task" [ 785.899318] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.907106] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52226bc7-ec48-bdf2-0fcc-106e4df64f0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.170086] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.243119] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.287682] env[61911]: DEBUG oslo_vmware.api [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Task: {'id': task-1250921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142502} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.287963] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.288192] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 786.288413] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.288538] env[61911]: INFO nova.compute.manager [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Took 1.48 seconds to destroy the instance on the hypervisor. [ 786.288804] env[61911]: DEBUG oslo.service.loopingcall [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.288999] env[61911]: DEBUG nova.compute.manager [-] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 786.289094] env[61911]: DEBUG nova.network.neutron [-] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.356206] env[61911]: DEBUG nova.compute.manager [req-00d7fb95-4d9f-43b7-8765-66bd66ab7479 req-a51d75fb-a2ae-4b58-bb9b-cf1bf18cba33 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Received event network-vif-plugged-6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 786.356456] env[61911]: DEBUG oslo_concurrency.lockutils [req-00d7fb95-4d9f-43b7-8765-66bd66ab7479 req-a51d75fb-a2ae-4b58-bb9b-cf1bf18cba33 service nova] Acquiring lock "28159e75-9fe9-44c7-b5c9-534902cecbef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.356633] env[61911]: DEBUG oslo_concurrency.lockutils [req-00d7fb95-4d9f-43b7-8765-66bd66ab7479 req-a51d75fb-a2ae-4b58-bb9b-cf1bf18cba33 service nova] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.356799] env[61911]: DEBUG oslo_concurrency.lockutils [req-00d7fb95-4d9f-43b7-8765-66bd66ab7479 req-a51d75fb-a2ae-4b58-bb9b-cf1bf18cba33 service nova] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.356965] env[61911]: DEBUG nova.compute.manager [req-00d7fb95-4d9f-43b7-8765-66bd66ab7479 req-a51d75fb-a2ae-4b58-bb9b-cf1bf18cba33 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] No waiting events found dispatching network-vif-plugged-6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 786.357871] env[61911]: WARNING nova.compute.manager [req-00d7fb95-4d9f-43b7-8765-66bd66ab7479 req-a51d75fb-a2ae-4b58-bb9b-cf1bf18cba33 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Received unexpected event network-vif-plugged-6531946d-4a56-42a8-9fcc-9ba3c928ad5d for instance with vm_state building and task_state spawning. [ 786.361868] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.382431] env[61911]: INFO nova.compute.manager [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Took 28.40 seconds to build instance. [ 786.411825] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52226bc7-ec48-bdf2-0fcc-106e4df64f0f, 'name': SearchDatastore_Task, 'duration_secs': 0.008691} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.412108] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.412359] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 267075d7-c3d8-4520-b7d9-67504097f823/267075d7-c3d8-4520-b7d9-67504097f823.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 786.412604] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f40110a1-46a6-40bb-a8df-027384deb550 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.419215] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 786.419215] env[61911]: value = "task-1250922" [ 786.419215] env[61911]: _type = "Task" [ 786.419215] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.426802] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.453059] env[61911]: DEBUG nova.network.neutron [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Successfully updated port: 6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 786.746100] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Releasing lock "refresh_cache-d4f17919-9cc9-4e19-88eb-38a978115975" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.746331] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61911) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 786.746520] env[61911]: DEBUG nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 786.746686] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.767765] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.809020] env[61911]: DEBUG nova.compute.manager [req-c9ca9d79-1149-49e2-bb95-8d99a886188e req-68095772-48f4-46ec-93c1-f12bafb6936c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Received event network-vif-deleted-14cd0e60-af95-417b-a8b0-6f143afa5e8b {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 786.809020] env[61911]: INFO nova.compute.manager [req-c9ca9d79-1149-49e2-bb95-8d99a886188e req-68095772-48f4-46ec-93c1-f12bafb6936c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Neutron deleted interface 14cd0e60-af95-417b-a8b0-6f143afa5e8b; detaching it from the instance and deleting it from the info cache [ 786.809020] env[61911]: DEBUG nova.network.neutron [req-c9ca9d79-1149-49e2-bb95-8d99a886188e req-68095772-48f4-46ec-93c1-f12bafb6936c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.884799] env[61911]: DEBUG oslo_concurrency.lockutils [None req-28ec21ea-a303-4b8f-8488-15d24f0cff81 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "66472f43-537d-4eb3-8d49-d40627a8809d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.872s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.937307] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250922, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440771} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.937479] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 267075d7-c3d8-4520-b7d9-67504097f823/267075d7-c3d8-4520-b7d9-67504097f823.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 786.937790] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 786.938356] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f1db62b-eba6-49d1-908f-8650a04b60cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.945171] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 786.945171] env[61911]: value = "task-1250923" [ 786.945171] env[61911]: _type = "Task" [ 786.945171] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.950695] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f2cd74-0f16-45e5-b096-6ea3c9a9dab4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.960051] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.960179] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.960335] env[61911]: DEBUG nova.network.neutron [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.965625] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.968588] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc26e8a-d076-464a-8ada-4294db591fb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.001460] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a039e4b-7d5a-4e39-9352-c629b7434aea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.008844] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a4f3e5-0c10-43df-973a-15c116bf06fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.022178] env[61911]: DEBUG nova.compute.provider_tree [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.228964] env[61911]: DEBUG nova.network.neutron [-] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.270674] env[61911]: DEBUG nova.network.neutron [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.309508] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6355cd89-419b-4724-8594-e4aa033c0b15 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.322175] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db30614c-8dbe-41ff-84f8-d96614770fc1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.345294] env[61911]: DEBUG nova.compute.manager [req-c9ca9d79-1149-49e2-bb95-8d99a886188e req-68095772-48f4-46ec-93c1-f12bafb6936c service nova] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Detach interface failed, port_id=14cd0e60-af95-417b-a8b0-6f143afa5e8b, reason: Instance 21723e9a-5c6d-4142-baba-25c59a5038ea could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 787.389658] env[61911]: DEBUG nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 787.456042] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066674} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.456368] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 787.457193] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95317a39-587e-4a19-9ab4-3e89814ddfae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.482984] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 267075d7-c3d8-4520-b7d9-67504097f823/267075d7-c3d8-4520-b7d9-67504097f823.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 787.483299] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f92e311-ab19-42bc-9a41-f94e70d64940 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.503207] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 787.503207] env[61911]: value = "task-1250924" [ 787.503207] env[61911]: _type = "Task" [ 787.503207] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.510514] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250924, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.524553] env[61911]: DEBUG nova.scheduler.client.report [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 787.537735] env[61911]: DEBUG nova.network.neutron [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.708237] env[61911]: DEBUG nova.network.neutron [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Updating instance_info_cache with network_info: [{"id": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "address": "fa:16:3e:8b:dc:1d", "network": {"id": "08a38fc2-219a-4060-849d-9cc2f226d818", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1686086958-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3fa4a98db1e4f39831a88402ec1ef6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6531946d-4a", "ovs_interfaceid": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.730547] env[61911]: INFO nova.compute.manager [-] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Took 1.44 seconds to deallocate network for instance. [ 787.774178] env[61911]: INFO nova.compute.manager [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] [instance: d4f17919-9cc9-4e19-88eb-38a978115975] Took 1.03 seconds to deallocate network for instance. [ 787.934418] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.015231] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250924, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.030557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.031236] env[61911]: DEBUG nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 788.034402] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.060s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.036036] env[61911]: INFO nova.compute.claims [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.211440] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Releasing lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.211975] env[61911]: DEBUG nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Instance network_info: |[{"id": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "address": "fa:16:3e:8b:dc:1d", "network": {"id": "08a38fc2-219a-4060-849d-9cc2f226d818", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1686086958-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3fa4a98db1e4f39831a88402ec1ef6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6531946d-4a", "ovs_interfaceid": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 788.212653] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:dc:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec763be6-4041-4651-8fd7-3820cf0ab86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6531946d-4a56-42a8-9fcc-9ba3c928ad5d', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.229345] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Creating folder: Project (b3fa4a98db1e4f39831a88402ec1ef6a). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.229873] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9c74150-3439-4619-8602-55b6a8f0ff57 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.237467] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.241780] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Created folder: Project (b3fa4a98db1e4f39831a88402ec1ef6a) in parent group-v269521. [ 788.242123] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Creating folder: Instances. Parent ref: group-v269549. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.242491] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c90ad51-5058-4673-ac24-890baa354c2e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.253552] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Created folder: Instances in parent group-v269549. [ 788.255418] env[61911]: DEBUG oslo.service.loopingcall [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.255418] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.255418] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-368938cf-9443-4df4-b1f3-98d358a38e55 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.284023] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.284023] env[61911]: value = "task-1250927" [ 788.284023] env[61911]: _type = "Task" [ 788.284023] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.291901] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250927, 'name': CreateVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.384462] env[61911]: DEBUG nova.compute.manager [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Received event network-changed-6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 788.384462] env[61911]: DEBUG nova.compute.manager [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Refreshing instance network info cache due to event network-changed-6531946d-4a56-42a8-9fcc-9ba3c928ad5d. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 788.384776] env[61911]: DEBUG oslo_concurrency.lockutils [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] Acquiring lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.384840] env[61911]: DEBUG oslo_concurrency.lockutils [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] Acquired lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.385043] env[61911]: DEBUG nova.network.neutron [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Refreshing network info cache for port 6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.517096] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250924, 'name': ReconfigVM_Task, 'duration_secs': 1.008197} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.517549] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 267075d7-c3d8-4520-b7d9-67504097f823/267075d7-c3d8-4520-b7d9-67504097f823.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 788.518925] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4c46830-2903-4e2a-8422-03e2b8529059 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.533827] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 788.533827] env[61911]: value = "task-1250928" [ 788.533827] env[61911]: _type = "Task" [ 788.533827] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.540657] env[61911]: DEBUG nova.compute.utils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.548288] env[61911]: DEBUG nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 788.548682] env[61911]: DEBUG nova.network.neutron [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.555389] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250928, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.672874] env[61911]: DEBUG nova.policy [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b034026eede4f5ca4757b552f7dca51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5be8ec1d59a14d618906ec32d0e134e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.794460] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250927, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.809896] env[61911]: INFO nova.scheduler.client.report [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Deleted allocations for instance d4f17919-9cc9-4e19-88eb-38a978115975 [ 788.954219] env[61911]: DEBUG nova.network.neutron [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Successfully created port: 5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.043903] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250928, 'name': Rename_Task, 'duration_secs': 0.160812} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.045289] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 789.045581] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75a5d9b9-e470-4d06-bd2e-554c212e910b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.052160] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 789.052160] env[61911]: value = "task-1250929" [ 789.052160] env[61911]: _type = "Task" [ 789.052160] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.057044] env[61911]: DEBUG nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 789.066991] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250929, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.296379] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250927, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.317843] env[61911]: DEBUG oslo_concurrency.lockutils [None req-11fff9f3-a906-46ea-942c-31d4a928bac8 tempest-ServersNegativeTestMultiTenantJSON-378668108 tempest-ServersNegativeTestMultiTenantJSON-378668108-project-member] Lock "d4f17919-9cc9-4e19-88eb-38a978115975" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.147s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.363822] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c337e4-3293-4d89-a2bf-302a6bbb5a48 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.373321] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3196118e-b5ea-4b44-adc3-8f6250f2582e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.406525] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840b8b9e-38ce-407e-84f4-d170c2b00ffc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.416524] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a31070-dead-4290-b4be-6bffba4b2b24 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.433275] env[61911]: DEBUG nova.compute.provider_tree [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.563223] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250929, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.607026] env[61911]: DEBUG nova.network.neutron [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Updated VIF entry in instance network info cache for port 6531946d-4a56-42a8-9fcc-9ba3c928ad5d. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 789.607434] env[61911]: DEBUG nova.network.neutron [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Updating instance_info_cache with network_info: [{"id": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "address": "fa:16:3e:8b:dc:1d", "network": {"id": "08a38fc2-219a-4060-849d-9cc2f226d818", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1686086958-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3fa4a98db1e4f39831a88402ec1ef6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6531946d-4a", "ovs_interfaceid": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.797057] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250927, 'name': CreateVM_Task, 'duration_secs': 1.439092} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.797247] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.797946] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.798126] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.798442] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.798696] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad52ce8f-eae6-4ec1-bbb4-1c5a4c34f347 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.803509] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 789.803509] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521fa081-bd73-cf9d-def9-0594446d35d4" [ 789.803509] env[61911]: _type = "Task" [ 789.803509] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.813023] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521fa081-bd73-cf9d-def9-0594446d35d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.822848] env[61911]: DEBUG nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 789.939089] env[61911]: DEBUG nova.scheduler.client.report [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 790.067020] env[61911]: DEBUG oslo_vmware.api [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1250929, 'name': PowerOnVM_Task, 'duration_secs': 0.675912} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.067020] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.067020] env[61911]: INFO nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Took 8.75 seconds to spawn the instance on the hypervisor. [ 790.067020] env[61911]: DEBUG nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 790.067020] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63e5ad2-5242-4359-8f4b-ab6f3e25b630 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.073877] env[61911]: DEBUG nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 790.101646] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 790.102085] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.102472] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 790.102746] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.103094] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 790.103320] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 790.103586] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 790.103943] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 790.104261] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 790.104537] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 790.104967] env[61911]: DEBUG nova.virt.hardware [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 790.105904] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c948fa86-69e7-48dd-9bca-c3859daf07b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.109781] env[61911]: DEBUG oslo_concurrency.lockutils [req-505116c8-f17f-49f8-bf5a-66f6eefa5b75 req-c8b9bb5e-afa3-45d3-ab47-98fe73f220fa service nova] Releasing lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.116135] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6421823d-ab99-4512-a561-bff66b49e074 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.320025] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521fa081-bd73-cf9d-def9-0594446d35d4, 'name': SearchDatastore_Task, 'duration_secs': 0.01131} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.320025] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.320025] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.320025] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.320294] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.320294] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.320294] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-774aa739-e23e-4db9-a32d-f4ac1090dbc3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.333515] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.334094] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.338124] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c10b0628-77cd-4005-9d19-b7d7b0752f6f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.347020] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 790.347020] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e70069-b1e6-0557-0596-8e52a1cb47a7" [ 790.347020] env[61911]: _type = "Task" [ 790.347020] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.355488] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e70069-b1e6-0557-0596-8e52a1cb47a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.357883] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.443909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.444738] env[61911]: DEBUG nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 790.448560] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.530s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.450036] env[61911]: INFO nova.compute.claims [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.589537] env[61911]: INFO nova.compute.manager [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Took 28.28 seconds to build instance. [ 790.697130] env[61911]: DEBUG nova.compute.manager [req-627ca35f-93a2-4fa2-a87c-13aadc920461 req-c69bea4a-bfb2-468b-aa2a-d4f114a637d0 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Received event network-vif-plugged-5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 790.697130] env[61911]: DEBUG oslo_concurrency.lockutils [req-627ca35f-93a2-4fa2-a87c-13aadc920461 req-c69bea4a-bfb2-468b-aa2a-d4f114a637d0 service nova] Acquiring lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.697130] env[61911]: DEBUG oslo_concurrency.lockutils [req-627ca35f-93a2-4fa2-a87c-13aadc920461 req-c69bea4a-bfb2-468b-aa2a-d4f114a637d0 service nova] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.697130] env[61911]: DEBUG oslo_concurrency.lockutils [req-627ca35f-93a2-4fa2-a87c-13aadc920461 req-c69bea4a-bfb2-468b-aa2a-d4f114a637d0 service nova] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.697498] env[61911]: DEBUG nova.compute.manager [req-627ca35f-93a2-4fa2-a87c-13aadc920461 req-c69bea4a-bfb2-468b-aa2a-d4f114a637d0 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] No waiting events found dispatching network-vif-plugged-5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 790.698304] env[61911]: WARNING nova.compute.manager [req-627ca35f-93a2-4fa2-a87c-13aadc920461 req-c69bea4a-bfb2-468b-aa2a-d4f114a637d0 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Received unexpected event network-vif-plugged-5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 for instance with vm_state building and task_state spawning. [ 790.837093] env[61911]: DEBUG nova.network.neutron [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Successfully updated port: 5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.862464] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e70069-b1e6-0557-0596-8e52a1cb47a7, 'name': SearchDatastore_Task, 'duration_secs': 0.009726} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.863472] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2ed7fec-19b7-4cf8-b382-a13899137919 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.870644] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 790.870644] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522f28b1-941b-ab6d-6124-3af3a77a38ce" [ 790.870644] env[61911]: _type = "Task" [ 790.870644] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.880031] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522f28b1-941b-ab6d-6124-3af3a77a38ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.960603] env[61911]: DEBUG nova.compute.utils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.963660] env[61911]: DEBUG nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 790.963883] env[61911]: DEBUG nova.network.neutron [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 791.029808] env[61911]: DEBUG nova.policy [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d850b9eeb164e7883e19ebe4d35b643', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e00325b35c074e92a5c80bf2a2df08dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.094595] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d004c56-d7c7-447e-b027-037c76773be4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "267075d7-c3d8-4520-b7d9-67504097f823" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.538s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.340608] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-8a363b99-517c-45ff-9e33-e9ceff3a08b0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.341147] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-8a363b99-517c-45ff-9e33-e9ceff3a08b0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.341460] env[61911]: DEBUG nova.network.neutron [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.384538] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522f28b1-941b-ab6d-6124-3af3a77a38ce, 'name': SearchDatastore_Task, 'duration_secs': 0.024389} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.384538] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.385397] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 28159e75-9fe9-44c7-b5c9-534902cecbef/28159e75-9fe9-44c7-b5c9-534902cecbef.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.385713] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5438f12b-b5a9-4f91-9646-4c17f45e5dbb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.395079] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 791.395079] env[61911]: value = "task-1250930" [ 791.395079] env[61911]: _type = "Task" [ 791.395079] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.404824] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.464239] env[61911]: DEBUG nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 791.496742] env[61911]: DEBUG nova.network.neutron [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Successfully created port: f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.602098] env[61911]: DEBUG nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 791.817174] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a88bd9-912e-4d5a-8b1a-848b42c19fb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.825395] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d669a293-d120-4b98-935a-f17074c9c956 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.860653] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180379da-5f37-4033-b8b2-667436cdc514 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.871030] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b871cbe1-58ae-452c-9db6-d008efb96794 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.891562] env[61911]: DEBUG nova.compute.provider_tree [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.906489] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250930, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.916811] env[61911]: DEBUG nova.network.neutron [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.134301] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.136573] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "613544bc-92b9-4c8f-8a7b-f02205f347ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.137016] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.161917] env[61911]: DEBUG nova.network.neutron [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Updating instance_info_cache with network_info: [{"id": "5bfc0a97-4a46-47d5-b55d-8d8a0374ff89", "address": "fa:16:3e:23:16:89", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bfc0a97-4a", "ovs_interfaceid": "5bfc0a97-4a46-47d5-b55d-8d8a0374ff89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.402295] env[61911]: DEBUG nova.scheduler.client.report [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 792.408519] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250930, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568386} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.409061] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 28159e75-9fe9-44c7-b5c9-534902cecbef/28159e75-9fe9-44c7-b5c9-534902cecbef.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.409305] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.409549] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7eaf0b91-5ca9-4bca-9211-7037267d3414 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.416113] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 792.416113] env[61911]: value = "task-1250931" [ 792.416113] env[61911]: _type = "Task" [ 792.416113] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.426193] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250931, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.483912] env[61911]: DEBUG nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 792.514643] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 792.514988] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.515324] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 792.515631] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.515745] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 792.515930] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 792.516192] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 792.516397] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 792.516767] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 792.516767] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 792.517017] env[61911]: DEBUG nova.virt.hardware [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 792.517911] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc885e2-f92d-49fd-87bc-3651bff56ea8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.527036] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b419fa7-2dca-486f-82aa-2b964e7ca3cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.664213] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-8a363b99-517c-45ff-9e33-e9ceff3a08b0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.664583] env[61911]: DEBUG nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Instance network_info: |[{"id": "5bfc0a97-4a46-47d5-b55d-8d8a0374ff89", "address": "fa:16:3e:23:16:89", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bfc0a97-4a", "ovs_interfaceid": "5bfc0a97-4a46-47d5-b55d-8d8a0374ff89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 792.665031] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:16:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bfc0a97-4a46-47d5-b55d-8d8a0374ff89', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.672944] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating folder: Project (5be8ec1d59a14d618906ec32d0e134e5). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.672944] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-12415c0e-b653-41cc-9674-15ce800d7b71 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.684622] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created folder: Project (5be8ec1d59a14d618906ec32d0e134e5) in parent group-v269521. [ 792.684873] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating folder: Instances. Parent ref: group-v269552. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.685218] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5149afb5-aaea-4fcd-a6ba-6b4b2a5329c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.695178] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created folder: Instances in parent group-v269552. [ 792.695434] env[61911]: DEBUG oslo.service.loopingcall [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.695638] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.697040] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69cff866-13a6-4900-a118-3809d168bc73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.719769] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.719769] env[61911]: value = "task-1250934" [ 792.719769] env[61911]: _type = "Task" [ 792.719769] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.731135] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250934, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.768200] env[61911]: DEBUG nova.compute.manager [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Received event network-changed-5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 792.768200] env[61911]: DEBUG nova.compute.manager [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Refreshing instance network info cache due to event network-changed-5bfc0a97-4a46-47d5-b55d-8d8a0374ff89. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 792.768200] env[61911]: DEBUG oslo_concurrency.lockutils [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] Acquiring lock "refresh_cache-8a363b99-517c-45ff-9e33-e9ceff3a08b0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.768200] env[61911]: DEBUG oslo_concurrency.lockutils [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] Acquired lock "refresh_cache-8a363b99-517c-45ff-9e33-e9ceff3a08b0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.768200] env[61911]: DEBUG nova.network.neutron [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Refreshing network info cache for port 5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.910788] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.911334] env[61911]: DEBUG nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 792.913961] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.225s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.915664] env[61911]: INFO nova.compute.claims [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.933479] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250931, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085189} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.933753] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.937962] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46356043-be4f-405f-8808-1243c148eec0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.962688] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 28159e75-9fe9-44c7-b5c9-534902cecbef/28159e75-9fe9-44c7-b5c9-534902cecbef.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.963703] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-959d2531-b4a5-45e6-b0f9-458524dd9ab7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.985980] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 792.985980] env[61911]: value = "task-1250935" [ 792.985980] env[61911]: _type = "Task" [ 792.985980] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.994172] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.232026] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250934, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.424509] env[61911]: DEBUG nova.compute.utils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.429562] env[61911]: DEBUG nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 793.429746] env[61911]: DEBUG nova.network.neutron [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 793.497674] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250935, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.503843] env[61911]: DEBUG nova.network.neutron [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Successfully updated port: f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.559511] env[61911]: DEBUG nova.policy [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d0056b50ef7429ea460ff4c1156d132', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf7a7083777a43a5bb9fc04584dcdd06', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.737663] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250934, 'name': CreateVM_Task, 'duration_secs': 0.726807} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.737960] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.738633] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.738813] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.739171] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.739736] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-633931ca-aa8f-4442-ba67-22a8110973f9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.744609] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 793.744609] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522deee2-4c31-c7c5-facf-2e6c72253aeb" [ 793.744609] env[61911]: _type = "Task" [ 793.744609] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.753254] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522deee2-4c31-c7c5-facf-2e6c72253aeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.866553] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "6e100d92-eb95-4b9b-b236-34ff94051811" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.866706] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "6e100d92-eb95-4b9b-b236-34ff94051811" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.913738] env[61911]: DEBUG nova.network.neutron [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Updated VIF entry in instance network info cache for port 5bfc0a97-4a46-47d5-b55d-8d8a0374ff89. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 793.914153] env[61911]: DEBUG nova.network.neutron [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Updating instance_info_cache with network_info: [{"id": "5bfc0a97-4a46-47d5-b55d-8d8a0374ff89", "address": "fa:16:3e:23:16:89", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bfc0a97-4a", "ovs_interfaceid": "5bfc0a97-4a46-47d5-b55d-8d8a0374ff89", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.931455] env[61911]: DEBUG nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 793.995399] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250935, 'name': ReconfigVM_Task, 'duration_secs': 0.584951} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.998059] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 28159e75-9fe9-44c7-b5c9-534902cecbef/28159e75-9fe9-44c7-b5c9-534902cecbef.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.998609] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-173dd08a-dd8b-402d-ab6f-665ef0a81391 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.006560] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.006834] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.006834] env[61911]: DEBUG nova.network.neutron [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.007960] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 794.007960] env[61911]: value = "task-1250936" [ 794.007960] env[61911]: _type = "Task" [ 794.007960] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.024141] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250936, 'name': Rename_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.221485] env[61911]: DEBUG nova.network.neutron [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Successfully created port: ddbc2d3e-91d8-4651-a95f-3f20c3ced622 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.241504] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb52957e-a27f-435a-84f0-dcc80f1bbe61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.251393] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2099746c-e633-47f9-acb4-02d004a1d912 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.258238] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522deee2-4c31-c7c5-facf-2e6c72253aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.017044} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.259678] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.259918] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.260181] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.260394] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.260614] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.261277] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85467b44-c937-4043-92f3-34d9491aed34 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.291961] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cbb57a-9304-41d2-a604-6a34da845f0a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.298743] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.298743] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 794.299565] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cb472b5-763d-428b-945a-fb8d1766b849 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.304382] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7b2f41-4061-4b92-ad7a-4538a046772b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.310927] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 794.310927] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bae533-bbf7-3d24-e73a-12faf8df54c6" [ 794.310927] env[61911]: _type = "Task" [ 794.310927] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.319171] env[61911]: DEBUG nova.compute.provider_tree [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.328183] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bae533-bbf7-3d24-e73a-12faf8df54c6, 'name': SearchDatastore_Task, 'duration_secs': 0.009413} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.329600] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eae3271c-2cde-4293-a2c6-c65505415f26 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.333846] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 794.333846] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527c4dca-dba8-e728-af7b-b5cd0a3bf84e" [ 794.333846] env[61911]: _type = "Task" [ 794.333846] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.342405] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527c4dca-dba8-e728-af7b-b5cd0a3bf84e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.417152] env[61911]: DEBUG oslo_concurrency.lockutils [req-cf38aabe-9af6-4bc2-844c-cc11d5433fd0 req-b549026f-dc40-48eb-98e1-0480dfb27903 service nova] Releasing lock "refresh_cache-8a363b99-517c-45ff-9e33-e9ceff3a08b0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.521108] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250936, 'name': Rename_Task, 'duration_secs': 0.16855} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.521108] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.521108] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62cfdbff-15ba-49dc-afa8-63f3baac7b18 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.530022] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 794.530022] env[61911]: value = "task-1250937" [ 794.530022] env[61911]: _type = "Task" [ 794.530022] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.539244] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250937, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.565637] env[61911]: DEBUG nova.network.neutron [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.825278] env[61911]: DEBUG nova.scheduler.client.report [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 794.830669] env[61911]: DEBUG nova.compute.manager [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received event network-vif-plugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 794.831015] env[61911]: DEBUG oslo_concurrency.lockutils [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.831301] env[61911]: DEBUG oslo_concurrency.lockutils [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.831631] env[61911]: DEBUG oslo_concurrency.lockutils [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.831752] env[61911]: DEBUG nova.compute.manager [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] No waiting events found dispatching network-vif-plugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 794.831975] env[61911]: WARNING nova.compute.manager [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received unexpected event network-vif-plugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c for instance with vm_state building and task_state spawning. [ 794.832191] env[61911]: DEBUG nova.compute.manager [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received event network-changed-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 794.832384] env[61911]: DEBUG nova.compute.manager [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Refreshing instance network info cache due to event network-changed-f6d1cfde-6743-4e5f-ac24-8b71f63c293c. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 794.832585] env[61911]: DEBUG oslo_concurrency.lockutils [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.845271] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527c4dca-dba8-e728-af7b-b5cd0a3bf84e, 'name': SearchDatastore_Task, 'duration_secs': 0.010519} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.846088] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.846382] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 794.846749] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9920c526-e2b7-46b2-94be-8ae193e41f61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.854029] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 794.854029] env[61911]: value = "task-1250938" [ 794.854029] env[61911]: _type = "Task" [ 794.854029] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.863037] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.863257] env[61911]: DEBUG nova.network.neutron [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.945130] env[61911]: DEBUG nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 794.971471] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 794.971714] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.972054] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 794.972260] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.972410] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 794.972555] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 794.972760] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 794.972935] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 794.974152] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 794.974416] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 794.974598] env[61911]: DEBUG nova.virt.hardware [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 794.975495] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771781bf-f43b-4f5d-8f28-211bf14c8143 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.983723] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9aaf0dc-536f-4037-8f51-4eeb9e121bc8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.037449] env[61911]: DEBUG oslo_vmware.api [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1250937, 'name': PowerOnVM_Task, 'duration_secs': 0.508981} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.037741] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.037949] env[61911]: INFO nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Took 9.24 seconds to spawn the instance on the hypervisor. [ 795.038143] env[61911]: DEBUG nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 795.038976] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27ea591-e800-410c-9c8e-bfdbd329befe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.335072] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.335509] env[61911]: DEBUG nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 795.338019] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 11.056s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.363266] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.365800] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.366097] env[61911]: DEBUG nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Instance network_info: |[{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 795.366369] env[61911]: DEBUG oslo_concurrency.lockutils [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.366540] env[61911]: DEBUG nova.network.neutron [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Refreshing network info cache for port f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 795.367607] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:e6:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6d1cfde-6743-4e5f-ac24-8b71f63c293c', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.375193] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating folder: Project (e00325b35c074e92a5c80bf2a2df08dd). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.379081] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74deeecc-0866-41cc-90d2-149a9664dc8c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.389776] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Created folder: Project (e00325b35c074e92a5c80bf2a2df08dd) in parent group-v269521. [ 795.389996] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating folder: Instances. Parent ref: group-v269555. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.390267] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05849c7e-a221-42c6-a954-c813f9351ee2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.401031] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Created folder: Instances in parent group-v269555. [ 795.401266] env[61911]: DEBUG oslo.service.loopingcall [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.401469] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 795.401692] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11224a5b-967c-4671-a4e7-9d388da06ed9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.423476] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.423476] env[61911]: value = "task-1250941" [ 795.423476] env[61911]: _type = "Task" [ 795.423476] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.431780] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250941, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.558594] env[61911]: INFO nova.compute.manager [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Took 28.98 seconds to build instance. [ 795.693224] env[61911]: DEBUG nova.network.neutron [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updated VIF entry in instance network info cache for port f6d1cfde-6743-4e5f-ac24-8b71f63c293c. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 795.693682] env[61911]: DEBUG nova.network.neutron [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.848273] env[61911]: DEBUG nova.compute.utils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.858024] env[61911]: DEBUG nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 795.858024] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 795.873208] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250938, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.00829} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.873208] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.873208] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.873208] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c4cac16-d6d7-48f7-b39c-484725f7f986 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.882828] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 795.882828] env[61911]: value = "task-1250942" [ 795.882828] env[61911]: _type = "Task" [ 795.882828] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.892259] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250942, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.933357] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250941, 'name': CreateVM_Task, 'duration_secs': 0.443917} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.933532] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 795.934773] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.934958] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.935359] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.935633] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59c678de-fee7-456c-8c4d-dc84a52c1f86 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.940349] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 795.940349] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e3f621-74d7-5b26-b91f-6675295d9ba0" [ 795.940349] env[61911]: _type = "Task" [ 795.940349] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.948586] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e3f621-74d7-5b26-b91f-6675295d9ba0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.989789] env[61911]: DEBUG nova.policy [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '142ced18076b4e1d80a56bae753dce9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbcd0cd6f3cf4c20ba385b874ba34fa1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.061732] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f8cb5df-20b5-445e-bed2-8e24cf664105 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.917s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.196419] env[61911]: DEBUG oslo_concurrency.lockutils [req-1d7752f3-23af-4ec1-8b85-fe2cd9ad152f req-aa17dcf8-945d-4acb-bdf8-c8b46702ee49 service nova] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.317407] env[61911]: DEBUG nova.compute.manager [req-25b71232-f165-4076-b12f-b5d3bf36bd1d req-7308b430-b065-42ce-acc5-06d77a4651ac service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Received event network-vif-plugged-ddbc2d3e-91d8-4651-a95f-3f20c3ced622 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 796.317619] env[61911]: DEBUG oslo_concurrency.lockutils [req-25b71232-f165-4076-b12f-b5d3bf36bd1d req-7308b430-b065-42ce-acc5-06d77a4651ac service nova] Acquiring lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.317826] env[61911]: DEBUG oslo_concurrency.lockutils [req-25b71232-f165-4076-b12f-b5d3bf36bd1d req-7308b430-b065-42ce-acc5-06d77a4651ac service nova] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.317991] env[61911]: DEBUG oslo_concurrency.lockutils [req-25b71232-f165-4076-b12f-b5d3bf36bd1d req-7308b430-b065-42ce-acc5-06d77a4651ac service nova] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.318501] env[61911]: DEBUG nova.compute.manager [req-25b71232-f165-4076-b12f-b5d3bf36bd1d req-7308b430-b065-42ce-acc5-06d77a4651ac service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] No waiting events found dispatching network-vif-plugged-ddbc2d3e-91d8-4651-a95f-3f20c3ced622 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 796.318767] env[61911]: WARNING nova.compute.manager [req-25b71232-f165-4076-b12f-b5d3bf36bd1d req-7308b430-b065-42ce-acc5-06d77a4651ac service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Received unexpected event network-vif-plugged-ddbc2d3e-91d8-4651-a95f-3f20c3ced622 for instance with vm_state building and task_state spawning. [ 796.360433] env[61911]: DEBUG nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 796.382945] env[61911]: WARNING nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 21723e9a-5c6d-4142-baba-25c59a5038ea is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 796.383202] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 66472f43-537d-4eb3-8d49-d40627a8809d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.383403] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 267075d7-c3d8-4520-b7d9-67504097f823 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.383589] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 28159e75-9fe9-44c7-b5c9-534902cecbef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.383719] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8a363b99-517c-45ff-9e33-e9ceff3a08b0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.383838] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8051e8ec-6d82-416c-858a-f0cdd00f869d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.384024] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 35cf6a24-3cc0-40bd-92da-798a56b4e2c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.384159] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 951e57ba-da09-426a-b500-e5459d346f64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 796.398384] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250942, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066604} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.399697] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.403759] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8856f6-c94f-49ab-abbf-e3885862eaf3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.429020] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.429655] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-667b045a-8a20-4149-a5c7-297f5b8c2f8d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.458416] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e3f621-74d7-5b26-b91f-6675295d9ba0, 'name': SearchDatastore_Task, 'duration_secs': 0.051999} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.462135] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.462135] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.462135] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.462135] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.462304] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.462304] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 796.462304] env[61911]: value = "task-1250943" [ 796.462304] env[61911]: _type = "Task" [ 796.462304] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.462304] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a6edb23-aa50-420e-b5c4-dccbf7591f12 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.472911] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.477667] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.477932] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.478656] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1baaf5e4-a53e-4775-b99a-ea717b90b202 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.483607] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 796.483607] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a35c53-e1f1-a376-2cbf-cdad9797bc35" [ 796.483607] env[61911]: _type = "Task" [ 796.483607] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.491796] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a35c53-e1f1-a376-2cbf-cdad9797bc35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.540690] env[61911]: DEBUG nova.network.neutron [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Successfully updated port: ddbc2d3e-91d8-4651-a95f-3f20c3ced622 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.567490] env[61911]: DEBUG nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 796.598697] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Successfully created port: fbb4bd3c-3d98-41d5-982a-a668a001e908 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.860239] env[61911]: DEBUG nova.compute.manager [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Received event network-changed-6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 796.860540] env[61911]: DEBUG nova.compute.manager [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Refreshing instance network info cache due to event network-changed-6531946d-4a56-42a8-9fcc-9ba3c928ad5d. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 796.860785] env[61911]: DEBUG oslo_concurrency.lockutils [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] Acquiring lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.861592] env[61911]: DEBUG oslo_concurrency.lockutils [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] Acquired lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.861592] env[61911]: DEBUG nova.network.neutron [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Refreshing network info cache for port 6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.874935] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Successfully created port: 11106334-7569-4e99-b691-e671ddf7fdaf {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.887744] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 28da8a83-a004-4baf-a744-bcc260032afd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.973825] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.993758] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a35c53-e1f1-a376-2cbf-cdad9797bc35, 'name': SearchDatastore_Task, 'duration_secs': 0.054052} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.994668] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d91a4c1-7c81-44a1-90a2-1a954d4810d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.000152] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 797.000152] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522eb9bc-25c4-803a-0b70-6fc779086cf3" [ 797.000152] env[61911]: _type = "Task" [ 797.000152] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.012604] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522eb9bc-25c4-803a-0b70-6fc779086cf3, 'name': SearchDatastore_Task, 'duration_secs': 0.009929} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.012858] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.013149] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 797.013394] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16f6ad9e-3c52-47f9-9cd9-3aa9a8d7e157 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.019891] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 797.019891] env[61911]: value = "task-1250944" [ 797.019891] env[61911]: _type = "Task" [ 797.019891] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.026862] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.042033] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "refresh_cache-35cf6a24-3cc0-40bd-92da-798a56b4e2c9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.042114] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "refresh_cache-35cf6a24-3cc0-40bd-92da-798a56b4e2c9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.043112] env[61911]: DEBUG nova.network.neutron [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.089415] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.369424] env[61911]: DEBUG nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 797.391861] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 43d18895-202c-4048-9435-b3484ffd4c07 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.404577] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 797.404817] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.405015] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 797.405253] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.405406] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 797.405556] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 797.405774] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 797.406110] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 797.406110] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 797.406291] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 797.406446] env[61911]: DEBUG nova.virt.hardware [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 797.407361] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3118863-14ca-41ac-a733-26a8a9d8a41a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.416753] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbd422a-b373-4f88-bbb3-0d4520a4cdf8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.474787] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250943, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.528819] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250944, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.593867] env[61911]: DEBUG nova.network.neutron [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.763417] env[61911]: DEBUG nova.network.neutron [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Updating instance_info_cache with network_info: [{"id": "ddbc2d3e-91d8-4651-a95f-3f20c3ced622", "address": "fa:16:3e:ab:cd:4e", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddbc2d3e-91", "ovs_interfaceid": "ddbc2d3e-91d8-4651-a95f-3f20c3ced622", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.821266] env[61911]: DEBUG nova.network.neutron [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Updated VIF entry in instance network info cache for port 6531946d-4a56-42a8-9fcc-9ba3c928ad5d. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 797.821671] env[61911]: DEBUG nova.network.neutron [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Updating instance_info_cache with network_info: [{"id": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "address": "fa:16:3e:8b:dc:1d", "network": {"id": "08a38fc2-219a-4060-849d-9cc2f226d818", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1686086958-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3fa4a98db1e4f39831a88402ec1ef6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6531946d-4a", "ovs_interfaceid": "6531946d-4a56-42a8-9fcc-9ba3c928ad5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.895372] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 31f0a4f7-a69d-4ec4-966d-93f402163b88 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 797.974616] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250943, 'name': ReconfigVM_Task, 'duration_secs': 1.03998} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.974873] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.975514] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91a24cc4-7c0e-4498-9536-5a518f0cdd79 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.982074] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 797.982074] env[61911]: value = "task-1250945" [ 797.982074] env[61911]: _type = "Task" [ 797.982074] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.988999] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250945, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.029813] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640757} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.030094] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 798.030311] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.030547] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfb22a92-05e4-4bef-a6b7-cc0d27aeb1b0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.036436] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 798.036436] env[61911]: value = "task-1250946" [ 798.036436] env[61911]: _type = "Task" [ 798.036436] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.043930] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.267327] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "refresh_cache-35cf6a24-3cc0-40bd-92da-798a56b4e2c9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.267677] env[61911]: DEBUG nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Instance network_info: |[{"id": "ddbc2d3e-91d8-4651-a95f-3f20c3ced622", "address": "fa:16:3e:ab:cd:4e", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddbc2d3e-91", "ovs_interfaceid": "ddbc2d3e-91d8-4651-a95f-3f20c3ced622", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 798.268242] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:cd:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddbc2d3e-91d8-4651-a95f-3f20c3ced622', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.276986] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating folder: Project (cf7a7083777a43a5bb9fc04584dcdd06). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.277314] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-665f9b74-2e09-4cc7-98ff-696c47dac718 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.287907] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Created folder: Project (cf7a7083777a43a5bb9fc04584dcdd06) in parent group-v269521. [ 798.288110] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating folder: Instances. Parent ref: group-v269558. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.288342] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f49991fa-be11-4c09-afc1-8923cba141c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.297827] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Created folder: Instances in parent group-v269558. [ 798.297993] env[61911]: DEBUG oslo.service.loopingcall [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.298213] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 798.298409] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4183db66-a35f-463a-be73-bc5404b80440 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.317245] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.317245] env[61911]: value = "task-1250949" [ 798.317245] env[61911]: _type = "Task" [ 798.317245] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.326037] env[61911]: DEBUG oslo_concurrency.lockutils [req-7904bc37-8871-4311-9203-fde9a982dbed req-037834dc-2316-4611-9529-2d36a8d34b76 service nova] Releasing lock "refresh_cache-28159e75-9fe9-44c7-b5c9-534902cecbef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.326037] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250949, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.400254] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance e9833b95-4162-42ba-87a4-d4cc790ac8e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 798.496915] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250945, 'name': Rename_Task, 'duration_secs': 0.143362} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.497163] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 798.497761] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-040833cb-5e03-4dee-92e6-a3f0bf82557b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.504473] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 798.504473] env[61911]: value = "task-1250950" [ 798.504473] env[61911]: _type = "Task" [ 798.504473] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.517874] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250950, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.519709] env[61911]: DEBUG nova.compute.manager [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Received event network-changed-ddbc2d3e-91d8-4651-a95f-3f20c3ced622 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 798.519916] env[61911]: DEBUG nova.compute.manager [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Refreshing instance network info cache due to event network-changed-ddbc2d3e-91d8-4651-a95f-3f20c3ced622. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 798.520179] env[61911]: DEBUG oslo_concurrency.lockutils [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] Acquiring lock "refresh_cache-35cf6a24-3cc0-40bd-92da-798a56b4e2c9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.520356] env[61911]: DEBUG oslo_concurrency.lockutils [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] Acquired lock "refresh_cache-35cf6a24-3cc0-40bd-92da-798a56b4e2c9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.520528] env[61911]: DEBUG nova.network.neutron [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Refreshing network info cache for port ddbc2d3e-91d8-4651-a95f-3f20c3ced622 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 798.547135] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059948} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.547544] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.548470] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8614432a-6d3f-40cd-89f0-a5e1c7d50c3c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.573287] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.573590] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9c7c51c-f2a4-4ea2-959e-ce86d3d74bdb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.593810] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 798.593810] env[61911]: value = "task-1250951" [ 798.593810] env[61911]: _type = "Task" [ 798.593810] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.602117] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.615641] env[61911]: DEBUG nova.compute.manager [req-8ff32cdb-ef0f-4c76-b068-4d2e2f4d0adc req-3b457f00-70bd-43c2-b3e1-8ddb95f1b92a service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received event network-vif-plugged-fbb4bd3c-3d98-41d5-982a-a668a001e908 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 798.615850] env[61911]: DEBUG oslo_concurrency.lockutils [req-8ff32cdb-ef0f-4c76-b068-4d2e2f4d0adc req-3b457f00-70bd-43c2-b3e1-8ddb95f1b92a service nova] Acquiring lock "951e57ba-da09-426a-b500-e5459d346f64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.615964] env[61911]: DEBUG oslo_concurrency.lockutils [req-8ff32cdb-ef0f-4c76-b068-4d2e2f4d0adc req-3b457f00-70bd-43c2-b3e1-8ddb95f1b92a service nova] Lock "951e57ba-da09-426a-b500-e5459d346f64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.616155] env[61911]: DEBUG oslo_concurrency.lockutils [req-8ff32cdb-ef0f-4c76-b068-4d2e2f4d0adc req-3b457f00-70bd-43c2-b3e1-8ddb95f1b92a service nova] Lock "951e57ba-da09-426a-b500-e5459d346f64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.616544] env[61911]: DEBUG nova.compute.manager [req-8ff32cdb-ef0f-4c76-b068-4d2e2f4d0adc req-3b457f00-70bd-43c2-b3e1-8ddb95f1b92a service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] No waiting events found dispatching network-vif-plugged-fbb4bd3c-3d98-41d5-982a-a668a001e908 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 798.616544] env[61911]: WARNING nova.compute.manager [req-8ff32cdb-ef0f-4c76-b068-4d2e2f4d0adc req-3b457f00-70bd-43c2-b3e1-8ddb95f1b92a service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received unexpected event network-vif-plugged-fbb4bd3c-3d98-41d5-982a-a668a001e908 for instance with vm_state building and task_state spawning. [ 798.828501] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250949, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.903700] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.016686] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250950, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.104643] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.129283] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Successfully updated port: fbb4bd3c-3d98-41d5-982a-a668a001e908 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.303075] env[61911]: DEBUG nova.network.neutron [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Updated VIF entry in instance network info cache for port ddbc2d3e-91d8-4651-a95f-3f20c3ced622. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.303470] env[61911]: DEBUG nova.network.neutron [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Updating instance_info_cache with network_info: [{"id": "ddbc2d3e-91d8-4651-a95f-3f20c3ced622", "address": "fa:16:3e:ab:cd:4e", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddbc2d3e-91", "ovs_interfaceid": "ddbc2d3e-91d8-4651-a95f-3f20c3ced622", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.328379] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250949, 'name': CreateVM_Task, 'duration_secs': 0.51324} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.328532] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.329201] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.329370] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.329683] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.329930] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8456be3-f9b4-4ad7-b588-0ed1e2e6306a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.334682] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 799.334682] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520be53f-45c5-47b4-441f-75a86f255eca" [ 799.334682] env[61911]: _type = "Task" [ 799.334682] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.342341] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520be53f-45c5-47b4-441f-75a86f255eca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.409169] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bafd58bc-8ae7-49a4-a039-31d328699010 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 799.515314] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250950, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.606469] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250951, 'name': ReconfigVM_Task, 'duration_secs': 0.878278} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.606679] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.607379] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77654862-b920-40f5-8206-fc48cdf584e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.613218] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 799.613218] env[61911]: value = "task-1250952" [ 799.613218] env[61911]: _type = "Task" [ 799.613218] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.622474] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250952, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.806910] env[61911]: DEBUG oslo_concurrency.lockutils [req-20851535-b5c0-46d8-8a8e-45ed434016d6 req-48674465-6b2b-476a-8514-4e3343d88007 service nova] Releasing lock "refresh_cache-35cf6a24-3cc0-40bd-92da-798a56b4e2c9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.845637] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520be53f-45c5-47b4-441f-75a86f255eca, 'name': SearchDatastore_Task, 'duration_secs': 0.00947} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.845980] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.848525] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.848525] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.848525] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.848525] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.848525] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02d36287-fa64-47f8-a506-3e13dd4eff11 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.856267] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.856525] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.857219] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac4322e1-ee12-4d40-b2ca-1e3ffc6fc7d8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.862654] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 799.862654] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a2d4fb-5ef9-7fe8-b941-e758d543dfc8" [ 799.862654] env[61911]: _type = "Task" [ 799.862654] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.870516] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a2d4fb-5ef9-7fe8-b941-e758d543dfc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.912560] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance c95707ca-7e32-42c0-914a-f0b178f07997 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.015633] env[61911]: DEBUG oslo_vmware.api [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250950, 'name': PowerOnVM_Task, 'duration_secs': 1.032738} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.015906] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 800.016151] env[61911]: INFO nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Took 9.94 seconds to spawn the instance on the hypervisor. [ 800.016298] env[61911]: DEBUG nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.017051] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec99ee6-9296-4566-b46d-49a082311f51 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.123850] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250952, 'name': Rename_Task, 'duration_secs': 0.156529} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.124284] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.125060] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3e7bfb4-3ca2-461a-bc98-2593aa6f73fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.132609] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 800.132609] env[61911]: value = "task-1250953" [ 800.132609] env[61911]: _type = "Task" [ 800.132609] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.142988] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250953, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.373461] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a2d4fb-5ef9-7fe8-b941-e758d543dfc8, 'name': SearchDatastore_Task, 'duration_secs': 0.010781} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.374250] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc561b5a-b796-4f5b-95f8-666eef538ba3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.379559] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 800.379559] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525710ef-d3b5-11c5-88bc-10b011d173e2" [ 800.379559] env[61911]: _type = "Task" [ 800.379559] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.387573] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525710ef-d3b5-11c5-88bc-10b011d173e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.415570] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 83616524-2530-4398-9019-148e0658720d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 800.534379] env[61911]: INFO nova.compute.manager [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Took 29.57 seconds to build instance. [ 800.543389] env[61911]: DEBUG nova.compute.manager [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received event network-changed-fbb4bd3c-3d98-41d5-982a-a668a001e908 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 800.543673] env[61911]: DEBUG nova.compute.manager [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Refreshing instance network info cache due to event network-changed-fbb4bd3c-3d98-41d5-982a-a668a001e908. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 800.543960] env[61911]: DEBUG oslo_concurrency.lockutils [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] Acquiring lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.544244] env[61911]: DEBUG oslo_concurrency.lockutils [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] Acquired lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.544420] env[61911]: DEBUG nova.network.neutron [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Refreshing network info cache for port fbb4bd3c-3d98-41d5-982a-a668a001e908 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.642376] env[61911]: DEBUG oslo_vmware.api [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1250953, 'name': PowerOnVM_Task, 'duration_secs': 0.490575} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.642658] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 800.642866] env[61911]: INFO nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Took 8.16 seconds to spawn the instance on the hypervisor. [ 800.643063] env[61911]: DEBUG nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.644033] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3df895-8801-4f51-86c9-4adc3c3ea0b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.891919] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525710ef-d3b5-11c5-88bc-10b011d173e2, 'name': SearchDatastore_Task, 'duration_secs': 0.009699} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.891919] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.892227] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 35cf6a24-3cc0-40bd-92da-798a56b4e2c9/35cf6a24-3cc0-40bd-92da-798a56b4e2c9.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.892544] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-041653ea-1425-42dd-b58b-b6c9d2d8d9dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.906192] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 800.906192] env[61911]: value = "task-1250954" [ 800.906192] env[61911]: _type = "Task" [ 800.906192] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.912016] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250954, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.919027] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 65d310c6-d438-4c3d-bd44-f3b51123fe93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.037068] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c2b0e259-8e0f-4814-90ac-44546b17c14b tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.317s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.140676] env[61911]: DEBUG nova.network.neutron [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.162451] env[61911]: INFO nova.compute.manager [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Took 28.21 seconds to build instance. [ 801.370184] env[61911]: DEBUG nova.network.neutron [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.416822] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250954, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466221} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.417459] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 35cf6a24-3cc0-40bd-92da-798a56b4e2c9/35cf6a24-3cc0-40bd-92da-798a56b4e2c9.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.417740] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.418056] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fd484ca-1546-45dd-8a7b-5d87defcd92d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.422416] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 613544bc-92b9-4c8f-8a7b-f02205f347ff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.425404] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 801.425404] env[61911]: value = "task-1250955" [ 801.425404] env[61911]: _type = "Task" [ 801.425404] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.436450] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250955, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.540378] env[61911]: DEBUG nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 801.557135] env[61911]: INFO nova.compute.manager [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Rebuilding instance [ 801.665569] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b4fc09-38d5-425d-a8e8-ce2f363732a6 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.404s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.695298] env[61911]: DEBUG nova.compute.manager [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 801.695298] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9359af8-8808-424a-8be2-a36d3065e4ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.698059] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Successfully updated port: 11106334-7569-4e99-b691-e671ddf7fdaf {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.873691] env[61911]: DEBUG oslo_concurrency.lockutils [req-71187a29-d097-43f8-983e-daefb9e28648 req-d5b5da08-4c19-43d6-88a3-1c0308d57173 service nova] Releasing lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.925879] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 6e100d92-eb95-4b9b-b236-34ff94051811 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 801.926468] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 801.926468] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 801.941849] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250955, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064688} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.942156] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.942970] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8337ecda-d0c6-4bbb-b499-03855f1ecada {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.946418] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Refreshing inventories for resource provider b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 801.973225] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 35cf6a24-3cc0-40bd-92da-798a56b4e2c9/35cf6a24-3cc0-40bd-92da-798a56b4e2c9.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.973225] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Updating ProviderTree inventory for provider b8a56394-897a-4e67-92ba-db46db6115e9 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 801.973225] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 801.975157] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49eb6c02-bcbe-4594-8c7f-3d4dd586865e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.995458] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 801.995458] env[61911]: value = "task-1250956" [ 801.995458] env[61911]: _type = "Task" [ 801.995458] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.003650] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250956, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.008115] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Refreshing aggregate associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, aggregates: None {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 802.010353] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.010647] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.028167] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Refreshing trait associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 802.061098] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.169580] env[61911]: DEBUG nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 802.200887] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.201045] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquired lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.201207] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.280271] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c47576-dfff-4a58-b0ef-d5121d9eaa7d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.288617] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77c6d46-b1e2-48e6-aead-c74e58d0ea3a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.318476] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed198f39-6485-4808-a6bc-3f16503499c0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.325707] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1444f61-a144-4dfb-84fb-acd33d765f69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.340648] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.505450] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250956, 'name': ReconfigVM_Task, 'duration_secs': 0.316984} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.506161] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 35cf6a24-3cc0-40bd-92da-798a56b4e2c9/35cf6a24-3cc0-40bd-92da-798a56b4e2c9.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.506639] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-31bd4b3f-d94a-4841-b195-3272fe750a73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.512511] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 802.512511] env[61911]: value = "task-1250957" [ 802.512511] env[61911]: _type = "Task" [ 802.512511] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.521902] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250957, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.577737] env[61911]: DEBUG nova.compute.manager [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received event network-vif-plugged-11106334-7569-4e99-b691-e671ddf7fdaf {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 802.577927] env[61911]: DEBUG oslo_concurrency.lockutils [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] Acquiring lock "951e57ba-da09-426a-b500-e5459d346f64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.578488] env[61911]: DEBUG oslo_concurrency.lockutils [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] Lock "951e57ba-da09-426a-b500-e5459d346f64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.578488] env[61911]: DEBUG oslo_concurrency.lockutils [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] Lock "951e57ba-da09-426a-b500-e5459d346f64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.578670] env[61911]: DEBUG nova.compute.manager [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] No waiting events found dispatching network-vif-plugged-11106334-7569-4e99-b691-e671ddf7fdaf {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 802.578898] env[61911]: WARNING nova.compute.manager [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received unexpected event network-vif-plugged-11106334-7569-4e99-b691-e671ddf7fdaf for instance with vm_state building and task_state spawning. [ 802.579098] env[61911]: DEBUG nova.compute.manager [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received event network-changed-11106334-7569-4e99-b691-e671ddf7fdaf {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 802.579264] env[61911]: DEBUG nova.compute.manager [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Refreshing instance network info cache due to event network-changed-11106334-7569-4e99-b691-e671ddf7fdaf. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 802.579431] env[61911]: DEBUG oslo_concurrency.lockutils [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] Acquiring lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.692832] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.712079] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 802.712399] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42168333-99bc-446b-99e3-70b4f253fd88 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.719194] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 802.719194] env[61911]: value = "task-1250958" [ 802.719194] env[61911]: _type = "Task" [ 802.719194] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.727661] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250958, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.769213] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.843617] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 803.031719] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250957, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.200360] env[61911]: DEBUG nova.network.neutron [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Updating instance_info_cache with network_info: [{"id": "fbb4bd3c-3d98-41d5-982a-a668a001e908", "address": "fa:16:3e:a8:a5:67", "network": {"id": "869636be-2b42-4253-af62-b2110c0cb283", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-857592256", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.76", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbcd0cd6f3cf4c20ba385b874ba34fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ac73085-7d7d-4ac3-863f-7b36a1b69aee", "external-id": "nsx-vlan-transportzone-289", "segmentation_id": 289, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbb4bd3c-3d", "ovs_interfaceid": "fbb4bd3c-3d98-41d5-982a-a668a001e908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "11106334-7569-4e99-b691-e671ddf7fdaf", "address": "fa:16:3e:93:b3:2e", "network": {"id": "338d9151-b857-4a8a-8d82-6d742771757d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2119394704", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.29", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dbcd0cd6f3cf4c20ba385b874ba34fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b8af79a-31d5-4d78-93d7-3919aa1d9186", "external-id": "nsx-vlan-transportzone-324", "segmentation_id": 324, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11106334-75", "ovs_interfaceid": "11106334-7569-4e99-b691-e671ddf7fdaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.229622] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250958, 'name': PowerOffVM_Task, 'duration_secs': 0.381509} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.229905] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 803.230195] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.230987] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8d24f7-f5a9-4a03-beed-65d5169f550d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.237962] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 803.238229] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82bd92e8-d423-48e2-ac3c-1ac7a9e0e3e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.297970] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 803.298197] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 803.298390] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleting the datastore file [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.298677] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-954b5290-394a-4e6e-a52a-79d2ecd360dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.304581] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 803.304581] env[61911]: value = "task-1250960" [ 803.304581] env[61911]: _type = "Task" [ 803.304581] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.312578] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.351964] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 803.352256] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.014s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.352562] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.543s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.354260] env[61911]: INFO nova.compute.claims [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.356889] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 803.357058] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Cleaning up deleted instances {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11608}} [ 803.523948] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250957, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.704074] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Releasing lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.704074] env[61911]: DEBUG nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Instance network_info: |[{"id": "fbb4bd3c-3d98-41d5-982a-a668a001e908", "address": "fa:16:3e:a8:a5:67", "network": {"id": "869636be-2b42-4253-af62-b2110c0cb283", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-857592256", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.76", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbcd0cd6f3cf4c20ba385b874ba34fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ac73085-7d7d-4ac3-863f-7b36a1b69aee", "external-id": "nsx-vlan-transportzone-289", "segmentation_id": 289, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbb4bd3c-3d", "ovs_interfaceid": "fbb4bd3c-3d98-41d5-982a-a668a001e908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "11106334-7569-4e99-b691-e671ddf7fdaf", "address": "fa:16:3e:93:b3:2e", "network": {"id": "338d9151-b857-4a8a-8d82-6d742771757d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2119394704", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.29", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dbcd0cd6f3cf4c20ba385b874ba34fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b8af79a-31d5-4d78-93d7-3919aa1d9186", "external-id": "nsx-vlan-transportzone-324", "segmentation_id": 324, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11106334-75", "ovs_interfaceid": "11106334-7569-4e99-b691-e671ddf7fdaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 803.704415] env[61911]: DEBUG oslo_concurrency.lockutils [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] Acquired lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.704458] env[61911]: DEBUG nova.network.neutron [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Refreshing network info cache for port 11106334-7569-4e99-b691-e671ddf7fdaf {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 803.705896] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:a5:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ac73085-7d7d-4ac3-863f-7b36a1b69aee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbb4bd3c-3d98-41d5-982a-a668a001e908', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:b3:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b8af79a-31d5-4d78-93d7-3919aa1d9186', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '11106334-7569-4e99-b691-e671ddf7fdaf', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.715443] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Creating folder: Project (dbcd0cd6f3cf4c20ba385b874ba34fa1). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.718851] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77258f66-ca26-47fb-b55d-c0cd556f1812 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.731192] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Created folder: Project (dbcd0cd6f3cf4c20ba385b874ba34fa1) in parent group-v269521. [ 803.731192] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Creating folder: Instances. Parent ref: group-v269561. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.731192] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19a2958c-c6d7-41ab-9a85-783f951959e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.739892] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Created folder: Instances in parent group-v269561. [ 803.740170] env[61911]: DEBUG oslo.service.loopingcall [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.740407] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 803.740707] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bee2275e-de7b-4ba4-9d67-d354710f27e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.762755] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.762755] env[61911]: value = "task-1250963" [ 803.762755] env[61911]: _type = "Task" [ 803.762755] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.770933] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250963, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.814649] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156112} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.815012] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 803.815115] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 803.815298] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 803.862359] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] There are 3 instances to clean {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11617}} [ 803.862621] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 7ad4d18c-cd25-4ea9-ac68-8457d0ca0b0e] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 804.025398] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250957, 'name': Rename_Task, 'duration_secs': 1.149137} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.025720] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.025957] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b42215e3-e03a-4b2e-aa11-c537e4c0baff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.034535] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 804.034535] env[61911]: value = "task-1250964" [ 804.034535] env[61911]: _type = "Task" [ 804.034535] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.043567] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250964, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.094794] env[61911]: DEBUG nova.network.neutron [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Updated VIF entry in instance network info cache for port 11106334-7569-4e99-b691-e671ddf7fdaf. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.095281] env[61911]: DEBUG nova.network.neutron [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Updating instance_info_cache with network_info: [{"id": "fbb4bd3c-3d98-41d5-982a-a668a001e908", "address": "fa:16:3e:a8:a5:67", "network": {"id": "869636be-2b42-4253-af62-b2110c0cb283", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-857592256", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.76", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbcd0cd6f3cf4c20ba385b874ba34fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ac73085-7d7d-4ac3-863f-7b36a1b69aee", "external-id": "nsx-vlan-transportzone-289", "segmentation_id": 289, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbb4bd3c-3d", "ovs_interfaceid": "fbb4bd3c-3d98-41d5-982a-a668a001e908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "11106334-7569-4e99-b691-e671ddf7fdaf", "address": "fa:16:3e:93:b3:2e", "network": {"id": "338d9151-b857-4a8a-8d82-6d742771757d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2119394704", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.29", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dbcd0cd6f3cf4c20ba385b874ba34fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b8af79a-31d5-4d78-93d7-3919aa1d9186", "external-id": "nsx-vlan-transportzone-324", "segmentation_id": 324, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11106334-75", "ovs_interfaceid": "11106334-7569-4e99-b691-e671ddf7fdaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.273082] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250963, 'name': CreateVM_Task, 'duration_secs': 0.386208} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.273268] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.274086] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.274286] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.274596] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.274851] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-164e47de-7c5e-41f3-8eb7-6f4d17cc51be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.279487] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 804.279487] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523ef3ea-eed7-992a-f014-cf44cafbc13a" [ 804.279487] env[61911]: _type = "Task" [ 804.279487] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.286841] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523ef3ea-eed7-992a-f014-cf44cafbc13a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.368769] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 90e2a5db-918a-4f66-a9c8-ef41dc4b855b] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 804.547167] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250964, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.598844] env[61911]: DEBUG oslo_concurrency.lockutils [req-64ae118a-9009-4e41-8b7f-838898ad4d4a req-6a7fd197-e46d-4f91-8096-d8ae047a735d service nova] Releasing lock "refresh_cache-951e57ba-da09-426a-b500-e5459d346f64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.613931] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5083dd-fe8a-4d3f-9935-14217b1fc153 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.621115] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405e7573-22d4-4307-85c2-e1cbfbe4f105 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.651834] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7d70d0-9c60-4078-a1f2-1971ab3efa76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.659234] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1220f759-8bd9-4aec-afe2-d273c906a84c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.673470] env[61911]: DEBUG nova.compute.provider_tree [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.790311] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523ef3ea-eed7-992a-f014-cf44cafbc13a, 'name': SearchDatastore_Task, 'duration_secs': 0.008247} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.790595] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.790863] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.791122] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.791273] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.791448] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.791696] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d40e6b2-a41e-496f-be6c-d006879378e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.799306] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.799514] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.800309] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68e2e32f-b3bc-4788-9b27-3a2381c4834a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.804970] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 804.804970] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a1a9ba-4733-9039-2d12-d0ecde772430" [ 804.804970] env[61911]: _type = "Task" [ 804.804970] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.812600] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a1a9ba-4733-9039-2d12-d0ecde772430, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.850164] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 804.850419] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.850580] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 804.850903] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.851133] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 804.851270] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 804.851479] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 804.851640] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 804.851806] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 804.851968] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 804.852184] env[61911]: DEBUG nova.virt.hardware [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 804.853151] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b191501c-1c24-4055-9a59-40181a335a0d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.860610] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a944f993-1f76-421f-84cd-f45c126c6369 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.873995] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 821d3f6f-e847-4ddb-ac00-ea55af302383] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 804.875739] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:16:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bfc0a97-4a46-47d5-b55d-8d8a0374ff89', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.883107] env[61911]: DEBUG oslo.service.loopingcall [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.883599] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 804.883828] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c57a4516-5722-41c7-964f-69cc474c7d27 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.903283] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.903283] env[61911]: value = "task-1250965" [ 804.903283] env[61911]: _type = "Task" [ 804.903283] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.911363] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250965, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.046776] env[61911]: DEBUG oslo_vmware.api [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250964, 'name': PowerOnVM_Task, 'duration_secs': 0.67243} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.047107] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.047336] env[61911]: INFO nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Took 10.10 seconds to spawn the instance on the hypervisor. [ 805.047532] env[61911]: DEBUG nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 805.048374] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87f308a-215c-4a7c-bb89-42373f155ab9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.176558] env[61911]: DEBUG nova.scheduler.client.report [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 805.315273] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a1a9ba-4733-9039-2d12-d0ecde772430, 'name': SearchDatastore_Task, 'duration_secs': 0.008089} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.316069] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17abbaf9-d0e0-4d73-a5ce-c64bf9998189 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.321691] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 805.321691] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5210b368-e6de-95c1-ed9f-72362dc03fcd" [ 805.321691] env[61911]: _type = "Task" [ 805.321691] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.329137] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5210b368-e6de-95c1-ed9f-72362dc03fcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.384981] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 805.385125] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Cleaning up deleted instances with incomplete migration {{(pid=61911) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11646}} [ 805.413513] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250965, 'name': CreateVM_Task, 'duration_secs': 0.281943} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.413677] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.414392] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.414562] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.414868] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.415129] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99f979b0-6259-4b7c-a8b3-3868e666950d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.419941] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 805.419941] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5279a635-a3b1-36a7-5bcc-c0c01b4a076d" [ 805.419941] env[61911]: _type = "Task" [ 805.419941] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.427744] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5279a635-a3b1-36a7-5bcc-c0c01b4a076d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.567049] env[61911]: INFO nova.compute.manager [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Took 30.67 seconds to build instance. [ 805.683479] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.684170] env[61911]: DEBUG nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 805.687401] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.326s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.688794] env[61911]: INFO nova.compute.claims [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 805.832271] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5210b368-e6de-95c1-ed9f-72362dc03fcd, 'name': SearchDatastore_Task, 'duration_secs': 0.008561} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.832747] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.832845] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 951e57ba-da09-426a-b500-e5459d346f64/951e57ba-da09-426a-b500-e5459d346f64.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.833060] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6a07964-d0bb-424c-a212-afec2fb70ff8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.839603] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 805.839603] env[61911]: value = "task-1250966" [ 805.839603] env[61911]: _type = "Task" [ 805.839603] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.847160] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250966, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.887766] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 805.929839] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5279a635-a3b1-36a7-5bcc-c0c01b4a076d, 'name': SearchDatastore_Task, 'duration_secs': 0.013579} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.930165] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.930424] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.930673] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.930833] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.931028] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.931295] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e48bfe02-fb2a-4934-8016-d2a20e5a1fa5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.939284] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.939442] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.940196] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e3ddc4d-f105-4012-a923-8392211b67b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.945452] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 805.945452] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52951289-36c0-712c-1514-e7b552c6180d" [ 805.945452] env[61911]: _type = "Task" [ 805.945452] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.953186] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52951289-36c0-712c-1514-e7b552c6180d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.069469] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b3ea6c8-e33b-4e7e-a4f0-ee8363fa9d31 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.827s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.193432] env[61911]: DEBUG nova.compute.utils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.197286] env[61911]: DEBUG nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 806.197481] env[61911]: DEBUG nova.network.neutron [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.300250] env[61911]: DEBUG nova.policy [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ecb17a7af574bf9bd6bcd33d8573afb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b09061b7653f4f05936988d99dcece50', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.351088] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250966, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443723} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.351376] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 951e57ba-da09-426a-b500-e5459d346f64/951e57ba-da09-426a-b500-e5459d346f64.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 806.351596] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.351840] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1af96ec-bf34-4b2c-a058-8f04116a499a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.357842] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 806.357842] env[61911]: value = "task-1250967" [ 806.357842] env[61911]: _type = "Task" [ 806.357842] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.366727] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250967, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.455203] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52951289-36c0-712c-1514-e7b552c6180d, 'name': SearchDatastore_Task, 'duration_secs': 0.008559} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.456029] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63a98aee-5ff0-4107-bde7-158f87ef2cc6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.460972] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 806.460972] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527176f2-6ecd-299f-7292-f21493256462" [ 806.460972] env[61911]: _type = "Task" [ 806.460972] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.468616] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527176f2-6ecd-299f-7292-f21493256462, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.572181] env[61911]: DEBUG nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 806.697782] env[61911]: DEBUG nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 806.711590] env[61911]: DEBUG nova.network.neutron [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Successfully created port: ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.870059] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250967, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.975542] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527176f2-6ecd-299f-7292-f21493256462, 'name': SearchDatastore_Task, 'duration_secs': 0.008853} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.975542] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.975668] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 806.975840] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4096e6d9-06fe-4cdc-8e10-8cdbee272a3b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.983077] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 806.983077] env[61911]: value = "task-1250968" [ 806.983077] env[61911]: _type = "Task" [ 806.983077] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.985296] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee98b237-268d-4cbd-a16b-b91ea41b7106 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.997955] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250968, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.998664] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49de12dd-730a-4767-8f72-3cd3d8bdc0f2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.033132] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166657f0-913b-44d0-b281-09105209a3ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.036149] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8bd7e1-8d18-4d86-9561-5bad21e40c1a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.045808] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c55f6d-df8f-4111-9700-eece95c31526 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.049549] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0fef0fc-451d-416c-b0de-212b4caa969e tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Suspending the VM {{(pid=61911) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 807.049795] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-aa8571b0-3eac-4fc1-b240-fa4c03642a73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.061456] env[61911]: DEBUG nova.compute.provider_tree [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.064332] env[61911]: DEBUG oslo_vmware.api [None req-a0fef0fc-451d-416c-b0de-212b4caa969e tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 807.064332] env[61911]: value = "task-1250969" [ 807.064332] env[61911]: _type = "Task" [ 807.064332] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.072087] env[61911]: DEBUG oslo_vmware.api [None req-a0fef0fc-451d-416c-b0de-212b4caa969e tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250969, 'name': SuspendVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.089777] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.371798] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250967, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.494054] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250968, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496899} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.494260] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.494561] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.494847] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea90da05-f155-4a1d-8a81-b946847db9c3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.500416] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 807.500416] env[61911]: value = "task-1250970" [ 807.500416] env[61911]: _type = "Task" [ 807.500416] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.508529] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.565361] env[61911]: DEBUG nova.scheduler.client.report [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 807.578957] env[61911]: DEBUG oslo_vmware.api [None req-a0fef0fc-451d-416c-b0de-212b4caa969e tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250969, 'name': SuspendVM_Task} progress is 54%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.711301] env[61911]: DEBUG nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 807.739022] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 807.739306] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.739470] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 807.739769] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.739995] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 807.740211] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 807.740485] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 807.740704] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 807.740944] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 807.741276] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 807.741401] env[61911]: DEBUG nova.virt.hardware [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 807.742428] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcd1e3a-9b60-4953-8e15-9a0643c38d94 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.752836] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2092c7be-aa74-4b6d-951c-3bd1b3ffa339 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.871579] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250967, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.009919] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060976} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.010218] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.011008] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87267c10-aee3-451b-a8fe-736ccf3b966d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.033098] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.033378] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66a4cefe-fb36-4e93-9560-fcfc60e8ac41 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.053291] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 808.053291] env[61911]: value = "task-1250971" [ 808.053291] env[61911]: _type = "Task" [ 808.053291] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.060311] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.074304] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.074927] env[61911]: DEBUG nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 808.077653] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.144s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.081046] env[61911]: INFO nova.compute.claims [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.086223] env[61911]: DEBUG oslo_vmware.api [None req-a0fef0fc-451d-416c-b0de-212b4caa969e tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250969, 'name': SuspendVM_Task, 'duration_secs': 0.727519} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.086460] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0fef0fc-451d-416c-b0de-212b4caa969e tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Suspended the VM {{(pid=61911) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 808.086635] env[61911]: DEBUG nova.compute.manager [None req-a0fef0fc-451d-416c-b0de-212b4caa969e tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 808.087427] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb7bb8f-21a7-448a-93f7-d1049ef17542 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.371828] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250967, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.562563] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250971, 'name': ReconfigVM_Task, 'duration_secs': 0.290561} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.562846] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0/8a363b99-517c-45ff-9e33-e9ceff3a08b0.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.563703] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-777d8da4-e78c-4637-a5ab-a8bb6ebc1e9c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.570318] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 808.570318] env[61911]: value = "task-1250972" [ 808.570318] env[61911]: _type = "Task" [ 808.570318] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.578187] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250972, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.583481] env[61911]: DEBUG nova.compute.utils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 808.589118] env[61911]: DEBUG nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 808.589118] env[61911]: DEBUG nova.network.neutron [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 808.642584] env[61911]: DEBUG nova.compute.manager [req-1150f89a-fbb9-404e-b0f4-b28c5e285edb req-a6b7c486-6e7b-42ce-bf73-f6d0967779f8 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Received event network-vif-plugged-ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 808.642886] env[61911]: DEBUG oslo_concurrency.lockutils [req-1150f89a-fbb9-404e-b0f4-b28c5e285edb req-a6b7c486-6e7b-42ce-bf73-f6d0967779f8 service nova] Acquiring lock "28da8a83-a004-4baf-a744-bcc260032afd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.643019] env[61911]: DEBUG oslo_concurrency.lockutils [req-1150f89a-fbb9-404e-b0f4-b28c5e285edb req-a6b7c486-6e7b-42ce-bf73-f6d0967779f8 service nova] Lock "28da8a83-a004-4baf-a744-bcc260032afd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.643198] env[61911]: DEBUG oslo_concurrency.lockutils [req-1150f89a-fbb9-404e-b0f4-b28c5e285edb req-a6b7c486-6e7b-42ce-bf73-f6d0967779f8 service nova] Lock "28da8a83-a004-4baf-a744-bcc260032afd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.643425] env[61911]: DEBUG nova.compute.manager [req-1150f89a-fbb9-404e-b0f4-b28c5e285edb req-a6b7c486-6e7b-42ce-bf73-f6d0967779f8 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] No waiting events found dispatching network-vif-plugged-ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 808.643600] env[61911]: WARNING nova.compute.manager [req-1150f89a-fbb9-404e-b0f4-b28c5e285edb req-a6b7c486-6e7b-42ce-bf73-f6d0967779f8 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Received unexpected event network-vif-plugged-ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 for instance with vm_state building and task_state spawning. [ 808.679603] env[61911]: DEBUG nova.policy [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18a64988ded64bb8b29f6c685bc82058', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a2306e0727146eaa8b8b250bca67966', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 808.772063] env[61911]: DEBUG nova.network.neutron [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Successfully updated port: ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.873313] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250967, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.066543] env[61911]: DEBUG nova.network.neutron [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Successfully created port: d7ec41d6-7384-43bc-86ac-98b776db99e8 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.079891] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250972, 'name': Rename_Task, 'duration_secs': 0.131661} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.080174] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.080363] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a150c74-369f-4cbb-9dea-be80e1a7f8e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.087029] env[61911]: DEBUG nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 809.089451] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 809.089451] env[61911]: value = "task-1250973" [ 809.089451] env[61911]: _type = "Task" [ 809.089451] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.100505] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250973, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.274737] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "refresh_cache-28da8a83-a004-4baf-a744-bcc260032afd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.274815] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquired lock "refresh_cache-28da8a83-a004-4baf-a744-bcc260032afd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.275038] env[61911]: DEBUG nova.network.neutron [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.378448] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250967, 'name': ExtendVirtualDisk_Task, 'duration_secs': 2.553232} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.378860] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.380084] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5ad037-d41d-4b58-9e67-c9d25bf26ca3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.419789] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 951e57ba-da09-426a-b500-e5459d346f64/951e57ba-da09-426a-b500-e5459d346f64.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.422692] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c351507-b8fe-4eb3-bcf2-a5798d4c8d89 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.437463] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424116f1-eb17-40e9-99a4-65abb56c69fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.449999] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7791e2f4-1c31-4511-b2f6-0531e0be0976 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.453373] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 809.453373] env[61911]: value = "task-1250974" [ 809.453373] env[61911]: _type = "Task" [ 809.453373] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.484271] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e22d6bb-7eed-40d7-a96d-c20262d874a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.490512] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250974, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.495952] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33e244f-16fa-4584-9738-895ac361a4aa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.511138] env[61911]: DEBUG nova.compute.provider_tree [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.606128] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250973, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.809528] env[61911]: DEBUG nova.network.neutron [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.973241] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250974, 'name': ReconfigVM_Task, 'duration_secs': 0.38059} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.973535] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 951e57ba-da09-426a-b500-e5459d346f64/951e57ba-da09-426a-b500-e5459d346f64.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.975538] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc1002f3-6e93-4bc1-a95a-879a7db1c6c8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.982200] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 809.982200] env[61911]: value = "task-1250975" [ 809.982200] env[61911]: _type = "Task" [ 809.982200] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.991453] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250975, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.013956] env[61911]: DEBUG nova.network.neutron [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Updating instance_info_cache with network_info: [{"id": "ea494ffc-7ea2-4dfd-84ed-1e401fe49c16", "address": "fa:16:3e:4d:49:21", "network": {"id": "7da127ea-75cf-4b23-b75e-891c44804c17", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-617016356-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b09061b7653f4f05936988d99dcece50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea494ffc-7e", "ovs_interfaceid": "ea494ffc-7ea2-4dfd-84ed-1e401fe49c16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.015814] env[61911]: DEBUG nova.scheduler.client.report [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 810.102059] env[61911]: DEBUG nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 810.107256] env[61911]: DEBUG oslo_vmware.api [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250973, 'name': PowerOnVM_Task, 'duration_secs': 0.723248} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.107592] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.107795] env[61911]: DEBUG nova.compute.manager [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 810.108735] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39384c87-9084-4499-a293-998b5453db32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.125252] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 810.125530] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.125710] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 810.125898] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.126070] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 810.126221] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 810.126432] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 810.126595] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 810.126810] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 810.126987] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 810.127217] env[61911]: DEBUG nova.virt.hardware [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 810.128081] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba53660-f66c-4851-b889-decaf6fed268 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.137822] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f33b6d-1d19-4f78-bf7f-6dcd6ff00ee1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.449831] env[61911]: DEBUG nova.compute.manager [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 810.450717] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37886d3e-5967-49aa-b9f9-860daa4636ca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.495752] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250975, 'name': Rename_Task, 'duration_secs': 0.179048} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.498216] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.498216] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6129d5a-13ac-4e4e-9d5e-38643676fb57 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.504748] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 810.504748] env[61911]: value = "task-1250976" [ 810.504748] env[61911]: _type = "Task" [ 810.504748] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.515095] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.519721] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Releasing lock "refresh_cache-28da8a83-a004-4baf-a744-bcc260032afd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.520038] env[61911]: DEBUG nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Instance network_info: |[{"id": "ea494ffc-7ea2-4dfd-84ed-1e401fe49c16", "address": "fa:16:3e:4d:49:21", "network": {"id": "7da127ea-75cf-4b23-b75e-891c44804c17", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-617016356-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b09061b7653f4f05936988d99dcece50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea494ffc-7e", "ovs_interfaceid": "ea494ffc-7ea2-4dfd-84ed-1e401fe49c16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 810.520776] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.521236] env[61911]: DEBUG nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 810.526018] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:49:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea494ffc-7ea2-4dfd-84ed-1e401fe49c16', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.532265] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Creating folder: Project (b09061b7653f4f05936988d99dcece50). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.532877] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.296s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.533119] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.535354] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.178s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.537026] env[61911]: INFO nova.compute.claims [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.539614] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f08ef4fe-3d6d-4c88-85ba-307290aea29f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.550990] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Created folder: Project (b09061b7653f4f05936988d99dcece50) in parent group-v269521. [ 810.551236] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Creating folder: Instances. Parent ref: group-v269565. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.551450] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-abef575a-502d-4752-a5af-06e3a28c0964 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.560446] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Created folder: Instances in parent group-v269565. [ 810.560726] env[61911]: DEBUG oslo.service.loopingcall [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.560881] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 810.561133] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bef90dcf-4e1d-4a55-b6b0-eec63e20d7c6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.576462] env[61911]: INFO nova.scheduler.client.report [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Deleted allocations for instance 21723e9a-5c6d-4142-baba-25c59a5038ea [ 810.582080] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.582080] env[61911]: value = "task-1250979" [ 810.582080] env[61911]: _type = "Task" [ 810.582080] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.591262] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250979, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.607799] env[61911]: DEBUG nova.network.neutron [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Successfully updated port: d7ec41d6-7384-43bc-86ac-98b776db99e8 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 810.639931] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.690499] env[61911]: DEBUG nova.compute.manager [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Received event network-changed-ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 810.690640] env[61911]: DEBUG nova.compute.manager [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Refreshing instance network info cache due to event network-changed-ea494ffc-7ea2-4dfd-84ed-1e401fe49c16. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 810.690879] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Acquiring lock "refresh_cache-28da8a83-a004-4baf-a744-bcc260032afd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.690985] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Acquired lock "refresh_cache-28da8a83-a004-4baf-a744-bcc260032afd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.691364] env[61911]: DEBUG nova.network.neutron [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Refreshing network info cache for port ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 810.963965] env[61911]: INFO nova.compute.manager [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] instance snapshotting [ 810.964208] env[61911]: WARNING nova.compute.manager [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 810.971029] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993f99b0-6032-4b01-aa6f-ca933f8c0a0a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.995012] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981dc44f-86fb-4677-ab6d-0b2a094c5df4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.016989] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250976, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.034932] env[61911]: DEBUG nova.compute.utils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.035950] env[61911]: DEBUG nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 811.036249] env[61911]: DEBUG nova.network.neutron [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 811.087861] env[61911]: DEBUG oslo_concurrency.lockutils [None req-77d769d9-b7fa-4019-864c-fe79d600ee47 tempest-ServerAddressesNegativeTestJSON-528356816 tempest-ServerAddressesNegativeTestJSON-528356816-project-member] Lock "21723e9a-5c6d-4142-baba-25c59a5038ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.815s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.090575] env[61911]: DEBUG nova.policy [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe50c877fcef489cb798fafe73caaf8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813b3181120245e594809d096a621675', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 811.099091] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250979, 'name': CreateVM_Task, 'duration_secs': 0.367813} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.100637] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.104753] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.105237] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.105554] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.106844] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b36c1682-451c-46c8-889a-e9ac97f7bd56 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.110480] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.110655] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquired lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.110724] env[61911]: DEBUG nova.network.neutron [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.113759] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 811.113759] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5244d7a5-1da8-7643-3cdf-bad0140a270e" [ 811.113759] env[61911]: _type = "Task" [ 811.113759] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.125079] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5244d7a5-1da8-7643-3cdf-bad0140a270e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.508978] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 811.508978] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d5e6fcb8-b64e-4567-aaba-66af12410f3a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.519154] env[61911]: DEBUG oslo_vmware.api [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250976, 'name': PowerOnVM_Task, 'duration_secs': 0.657602} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.520987] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.521713] env[61911]: INFO nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Took 14.15 seconds to spawn the instance on the hypervisor. [ 811.522459] env[61911]: DEBUG nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 811.523584] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 811.523584] env[61911]: value = "task-1250980" [ 811.523584] env[61911]: _type = "Task" [ 811.523584] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.524063] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d084de5-0dcf-4075-91fc-a297388de94d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.547746] env[61911]: DEBUG nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 811.552381] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250980, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.610180] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.610180] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.610180] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.610180] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.610180] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.618472] env[61911]: INFO nova.compute.manager [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Terminating instance [ 811.642134] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5244d7a5-1da8-7643-3cdf-bad0140a270e, 'name': SearchDatastore_Task, 'duration_secs': 0.012244} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.643566] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.643566] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.643566] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.643668] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.643896] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.649113] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e00a7a60-ca75-4264-9858-5820f458ffe0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.661609] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.661609] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.662224] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be0b0814-2dd7-4b9c-82aa-0c3c3ff95b08 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.672053] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 811.672053] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5262c7b8-1f56-75aa-7484-b11141b320d6" [ 811.672053] env[61911]: _type = "Task" [ 811.672053] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.672053] env[61911]: DEBUG nova.network.neutron [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Updated VIF entry in instance network info cache for port ea494ffc-7ea2-4dfd-84ed-1e401fe49c16. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 811.672053] env[61911]: DEBUG nova.network.neutron [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Updating instance_info_cache with network_info: [{"id": "ea494ffc-7ea2-4dfd-84ed-1e401fe49c16", "address": "fa:16:3e:4d:49:21", "network": {"id": "7da127ea-75cf-4b23-b75e-891c44804c17", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-617016356-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b09061b7653f4f05936988d99dcece50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea494ffc-7e", "ovs_interfaceid": "ea494ffc-7ea2-4dfd-84ed-1e401fe49c16", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.685308] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5262c7b8-1f56-75aa-7484-b11141b320d6, 'name': SearchDatastore_Task, 'duration_secs': 0.009522} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.687114] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-125d0a3f-7e9e-48a4-9667-57f6034d58fb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.694017] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 811.694017] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528a0ded-598f-2911-397d-6aa9dc84c780" [ 811.694017] env[61911]: _type = "Task" [ 811.694017] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.709380] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528a0ded-598f-2911-397d-6aa9dc84c780, 'name': SearchDatastore_Task, 'duration_secs': 0.008715} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.709706] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.710059] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 28da8a83-a004-4baf-a744-bcc260032afd/28da8a83-a004-4baf-a744-bcc260032afd.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.710403] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e053ee99-0036-49e0-bfc0-5ff15bb6a19c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.719400] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 811.719400] env[61911]: value = "task-1250981" [ 811.719400] env[61911]: _type = "Task" [ 811.719400] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.730177] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.733792] env[61911]: DEBUG nova.network.neutron [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.766841] env[61911]: DEBUG nova.network.neutron [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Successfully created port: 8d69ab4c-3b99-412f-8985-3d8a9210f776 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 811.934031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ed53cd-aaa2-4e33-9307-7a0f9111ddf7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.946507] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a95def6-20b0-4065-a49e-2c9b57d26fcd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.979781] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884d3089-91a8-4031-8605-76e13a3fa4e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.988698] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa6dc86-1ff1-47b1-85ad-08258a146382 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.004272] env[61911]: DEBUG nova.compute.provider_tree [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.006620] env[61911]: DEBUG nova.network.neutron [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Updating instance_info_cache with network_info: [{"id": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "address": "fa:16:3e:bc:60:07", "network": {"id": "d54e4625-67a5-4f4e-92cd-230f95420309", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2030291444-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a2306e0727146eaa8b8b250bca67966", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7ec41d6-73", "ovs_interfaceid": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.041166] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250980, 'name': CreateSnapshot_Task, 'duration_secs': 0.515968} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.041345] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 812.042189] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd021cf-c2b1-4fa1-8dda-6b2257ddca23 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.068064] env[61911]: INFO nova.compute.manager [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Took 30.40 seconds to build instance. [ 812.131900] env[61911]: DEBUG nova.compute.manager [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 812.132228] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 812.133329] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b30d933-b191-4344-bd3b-8388e44c7dea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.144640] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.145056] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-718cfe80-d336-4d64-b819-f3d471b9a207 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.154641] env[61911]: DEBUG oslo_vmware.api [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 812.154641] env[61911]: value = "task-1250982" [ 812.154641] env[61911]: _type = "Task" [ 812.154641] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.166277] env[61911]: DEBUG oslo_vmware.api [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250982, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.176727] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Releasing lock "refresh_cache-28da8a83-a004-4baf-a744-bcc260032afd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.177196] env[61911]: DEBUG nova.compute.manager [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Received event network-vif-plugged-d7ec41d6-7384-43bc-86ac-98b776db99e8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 812.177306] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Acquiring lock "43d18895-202c-4048-9435-b3484ffd4c07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.177533] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Lock "43d18895-202c-4048-9435-b3484ffd4c07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.177731] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Lock "43d18895-202c-4048-9435-b3484ffd4c07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.177911] env[61911]: DEBUG nova.compute.manager [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] No waiting events found dispatching network-vif-plugged-d7ec41d6-7384-43bc-86ac-98b776db99e8 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 812.178114] env[61911]: WARNING nova.compute.manager [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Received unexpected event network-vif-plugged-d7ec41d6-7384-43bc-86ac-98b776db99e8 for instance with vm_state building and task_state spawning. [ 812.179591] env[61911]: DEBUG nova.compute.manager [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Received event network-changed-d7ec41d6-7384-43bc-86ac-98b776db99e8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 812.179591] env[61911]: DEBUG nova.compute.manager [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Refreshing instance network info cache due to event network-changed-d7ec41d6-7384-43bc-86ac-98b776db99e8. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 812.179591] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Acquiring lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.230141] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250981, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.395179] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "9230cd85-960d-427c-b69c-d19ac2bcfe35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.395179] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.509051] env[61911]: DEBUG nova.scheduler.client.report [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 812.512938] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Releasing lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.513234] env[61911]: DEBUG nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Instance network_info: |[{"id": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "address": "fa:16:3e:bc:60:07", "network": {"id": "d54e4625-67a5-4f4e-92cd-230f95420309", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2030291444-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a2306e0727146eaa8b8b250bca67966", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7ec41d6-73", "ovs_interfaceid": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 812.513707] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Acquired lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.513887] env[61911]: DEBUG nova.network.neutron [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Refreshing network info cache for port d7ec41d6-7384-43bc-86ac-98b776db99e8 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.514867] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:60:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60bdba1a-14cf-46b2-9d8b-aeaf4d80c815', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd7ec41d6-7384-43bc-86ac-98b776db99e8', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.523273] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Creating folder: Project (5a2306e0727146eaa8b8b250bca67966). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.525052] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9cfebe4-81d7-4707-8e70-9391d9207d90 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.536206] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Created folder: Project (5a2306e0727146eaa8b8b250bca67966) in parent group-v269521. [ 812.536410] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Creating folder: Instances. Parent ref: group-v269569. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.536643] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e824587e-2112-4877-afdf-4c473d38cadb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.545053] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Created folder: Instances in parent group-v269569. [ 812.545291] env[61911]: DEBUG oslo.service.loopingcall [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.545475] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 812.545675] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-febf8113-b3c3-4431-99b3-c66532e2f77c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.568350] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 812.570010] env[61911]: DEBUG nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 812.571817] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-07f96ad3-4715-47cc-bd1a-674a38873d3a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.575230] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fbde6fbc-6ee9-412c-bdbb-b2ddc191dc7a tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "951e57ba-da09-426a-b500-e5459d346f64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.765s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.580031] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 812.580031] env[61911]: value = "task-1250985" [ 812.580031] env[61911]: _type = "Task" [ 812.580031] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.581370] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 812.581370] env[61911]: value = "task-1250986" [ 812.581370] env[61911]: _type = "Task" [ 812.581370] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.584448] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "951e57ba-da09-426a-b500-e5459d346f64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.585239] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "951e57ba-da09-426a-b500-e5459d346f64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.585239] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "951e57ba-da09-426a-b500-e5459d346f64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.585239] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "951e57ba-da09-426a-b500-e5459d346f64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.585239] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "951e57ba-da09-426a-b500-e5459d346f64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.590120] env[61911]: INFO nova.compute.manager [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Terminating instance [ 812.605660] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250986, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.606000] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250985, 'name': CreateVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.608306] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 812.608536] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.608696] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 812.608883] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.609043] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 812.609198] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 812.609407] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 812.609629] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 812.609827] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 812.610020] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 812.611027] env[61911]: DEBUG nova.virt.hardware [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 812.611429] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd2d72e-1fe2-4727-83d5-e5bd6e3c722d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.619753] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d2d66e-84d9-467e-bcf0-b9a05bdc0a0c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.665028] env[61911]: DEBUG oslo_vmware.api [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250982, 'name': PowerOffVM_Task, 'duration_secs': 0.332426} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.665232] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 812.665462] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 812.666032] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5aae80b0-06ac-42e5-8a1f-144166bc0853 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.730436] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546832} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.730738] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 28da8a83-a004-4baf-a744-bcc260032afd/28da8a83-a004-4baf-a744-bcc260032afd.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.731038] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.731800] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec1f1a74-3d1c-4ec7-a9c8-f326425867f4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.738023] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 812.738023] env[61911]: value = "task-1250988" [ 812.738023] env[61911]: _type = "Task" [ 812.738023] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.746649] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.750917] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 812.751309] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 812.751604] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleting the datastore file [datastore2] 8a363b99-517c-45ff-9e33-e9ceff3a08b0 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.751920] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ad005a2-058c-426b-847a-8c09da20c189 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.759620] env[61911]: DEBUG oslo_vmware.api [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 812.759620] env[61911]: value = "task-1250989" [ 812.759620] env[61911]: _type = "Task" [ 812.759620] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.769331] env[61911]: DEBUG oslo_vmware.api [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250989, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.014878] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.015520] env[61911]: DEBUG nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 813.018812] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.885s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.020544] env[61911]: INFO nova.compute.claims [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.078522] env[61911]: DEBUG nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 813.095202] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250985, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.098900] env[61911]: DEBUG nova.compute.manager [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 813.099232] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 813.099572] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250986, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.100657] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61095b3-e958-42e6-a796-a8c889c437f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.109255] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 813.109255] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d74a4b1-784d-4a09-97a3-34fc02acbdad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.119212] env[61911]: DEBUG oslo_vmware.api [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 813.119212] env[61911]: value = "task-1250990" [ 813.119212] env[61911]: _type = "Task" [ 813.119212] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.129182] env[61911]: DEBUG oslo_vmware.api [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.248544] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1801} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.248865] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.249616] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbbb430-5bb3-4cda-bbcb-ecab3dfe55b6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.253130] env[61911]: DEBUG nova.network.neutron [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Updated VIF entry in instance network info cache for port d7ec41d6-7384-43bc-86ac-98b776db99e8. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.253498] env[61911]: DEBUG nova.network.neutron [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Updating instance_info_cache with network_info: [{"id": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "address": "fa:16:3e:bc:60:07", "network": {"id": "d54e4625-67a5-4f4e-92cd-230f95420309", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2030291444-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a2306e0727146eaa8b8b250bca67966", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7ec41d6-73", "ovs_interfaceid": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.275137] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 28da8a83-a004-4baf-a744-bcc260032afd/28da8a83-a004-4baf-a744-bcc260032afd.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.276591] env[61911]: DEBUG oslo_concurrency.lockutils [req-64723a6c-2fbf-4c15-bb4b-e29ad2e11624 req-5ccdc2d4-e69b-460f-a0f9-f35454bbadc6 service nova] Releasing lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.279941] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6911bf20-88a5-42eb-83c3-9a0acc7386c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.302493] env[61911]: DEBUG oslo_vmware.api [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1250989, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.415819} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.303982] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.304225] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 813.304420] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 813.304628] env[61911]: INFO nova.compute.manager [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Took 1.17 seconds to destroy the instance on the hypervisor. [ 813.304905] env[61911]: DEBUG oslo.service.loopingcall [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.305564] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 813.305564] env[61911]: value = "task-1250991" [ 813.305564] env[61911]: _type = "Task" [ 813.305564] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.306052] env[61911]: DEBUG nova.compute.manager [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 813.306052] env[61911]: DEBUG nova.network.neutron [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.317474] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250991, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.525151] env[61911]: DEBUG nova.compute.utils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.530051] env[61911]: DEBUG nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 813.530051] env[61911]: DEBUG nova.network.neutron [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.599388] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250985, 'name': CreateVM_Task, 'duration_secs': 0.548689} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.602530] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 813.602793] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250986, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.603443] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.603670] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.603902] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 813.604452] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69faf368-f558-4040-a189-624537714f3e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.609014] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 813.609014] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268f38f-3e0b-5a84-c530-b55531657f0c" [ 813.609014] env[61911]: _type = "Task" [ 813.609014] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.609954] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.618620] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268f38f-3e0b-5a84-c530-b55531657f0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.627212] env[61911]: DEBUG oslo_vmware.api [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250990, 'name': PowerOffVM_Task, 'duration_secs': 0.205489} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.627456] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 813.627624] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 813.627855] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f665098-90f4-4dbd-bd65-6cff0445d1d1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.654331] env[61911]: DEBUG nova.policy [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ee96b52414148b7bff1d067595848b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f40f512384d4fedb75bfc53f3195243', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.660090] env[61911]: DEBUG nova.compute.manager [req-2a49bd88-1c80-4b33-bb11-3b31c11d72ab req-b6dd74c3-26d6-49c5-b7d3-40405b7b3bf7 service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Received event network-vif-plugged-8d69ab4c-3b99-412f-8985-3d8a9210f776 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 813.660306] env[61911]: DEBUG oslo_concurrency.lockutils [req-2a49bd88-1c80-4b33-bb11-3b31c11d72ab req-b6dd74c3-26d6-49c5-b7d3-40405b7b3bf7 service nova] Acquiring lock "31f0a4f7-a69d-4ec4-966d-93f402163b88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.660513] env[61911]: DEBUG oslo_concurrency.lockutils [req-2a49bd88-1c80-4b33-bb11-3b31c11d72ab req-b6dd74c3-26d6-49c5-b7d3-40405b7b3bf7 service nova] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.660682] env[61911]: DEBUG oslo_concurrency.lockutils [req-2a49bd88-1c80-4b33-bb11-3b31c11d72ab req-b6dd74c3-26d6-49c5-b7d3-40405b7b3bf7 service nova] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.660870] env[61911]: DEBUG nova.compute.manager [req-2a49bd88-1c80-4b33-bb11-3b31c11d72ab req-b6dd74c3-26d6-49c5-b7d3-40405b7b3bf7 service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] No waiting events found dispatching network-vif-plugged-8d69ab4c-3b99-412f-8985-3d8a9210f776 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 813.661139] env[61911]: WARNING nova.compute.manager [req-2a49bd88-1c80-4b33-bb11-3b31c11d72ab req-b6dd74c3-26d6-49c5-b7d3-40405b7b3bf7 service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Received unexpected event network-vif-plugged-8d69ab4c-3b99-412f-8985-3d8a9210f776 for instance with vm_state building and task_state spawning. [ 813.682483] env[61911]: DEBUG nova.compute.manager [req-5db89a3a-9ad2-41d7-a70e-986f5f760a74 req-29b2f41f-f29f-4ed5-8f15-7e978e653b8a service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Received event network-vif-deleted-5bfc0a97-4a46-47d5-b55d-8d8a0374ff89 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 813.682692] env[61911]: INFO nova.compute.manager [req-5db89a3a-9ad2-41d7-a70e-986f5f760a74 req-29b2f41f-f29f-4ed5-8f15-7e978e653b8a service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Neutron deleted interface 5bfc0a97-4a46-47d5-b55d-8d8a0374ff89; detaching it from the instance and deleting it from the info cache [ 813.682866] env[61911]: DEBUG nova.network.neutron [req-5db89a3a-9ad2-41d7-a70e-986f5f760a74 req-29b2f41f-f29f-4ed5-8f15-7e978e653b8a service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.719638] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 813.719948] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 813.720180] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Deleting the datastore file [datastore1] 951e57ba-da09-426a-b500-e5459d346f64 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.720437] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42197031-4a44-4202-973b-d6fb2e941603 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.726695] env[61911]: DEBUG oslo_vmware.api [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for the task: (returnval){ [ 813.726695] env[61911]: value = "task-1250993" [ 813.726695] env[61911]: _type = "Task" [ 813.726695] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.735098] env[61911]: DEBUG oslo_vmware.api [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.820095] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250991, 'name': ReconfigVM_Task, 'duration_secs': 0.307068} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.820095] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 28da8a83-a004-4baf-a744-bcc260032afd/28da8a83-a004-4baf-a744-bcc260032afd.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.820095] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77e1350b-ad58-4b01-ab70-53677e832263 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.826059] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 813.826059] env[61911]: value = "task-1250994" [ 813.826059] env[61911]: _type = "Task" [ 813.826059] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.834948] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250994, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.847145] env[61911]: DEBUG nova.network.neutron [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Successfully updated port: 8d69ab4c-3b99-412f-8985-3d8a9210f776 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.032619] env[61911]: DEBUG nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 814.076648] env[61911]: DEBUG nova.network.neutron [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.106645] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1250986, 'name': CloneVM_Task, 'duration_secs': 1.362683} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.106645] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Created linked-clone VM from snapshot [ 814.107236] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e92884-f76a-4672-86fc-81619a6e5752 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.120954] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268f38f-3e0b-5a84-c530-b55531657f0c, 'name': SearchDatastore_Task, 'duration_secs': 0.010412} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.124555] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.124826] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.125070] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.125249] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.125387] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.125688] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Uploading image 1dc41913-75eb-4381-b0aa-d373205152d8 {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 814.127903] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-828902a0-1a1d-4a23-b2a5-b6c5dc6d9209 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.139535] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.140393] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 814.140672] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98f532cb-4e8b-40fe-9507-44a75dfc48eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.151777] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 814.151777] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523fcf4f-21d9-958e-e015-8e03f10ac777" [ 814.151777] env[61911]: _type = "Task" [ 814.151777] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.157854] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 814.157854] env[61911]: value = "vm-269572" [ 814.157854] env[61911]: _type = "VirtualMachine" [ 814.157854] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 814.158137] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-70221783-ca1d-4d76-a694-d43aadf13354 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.166943] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523fcf4f-21d9-958e-e015-8e03f10ac777, 'name': SearchDatastore_Task, 'duration_secs': 0.0092} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.168766] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lease: (returnval){ [ 814.168766] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5266358b-ce95-d68b-2d99-e313d43515c3" [ 814.168766] env[61911]: _type = "HttpNfcLease" [ 814.168766] env[61911]: } obtained for exporting VM: (result){ [ 814.168766] env[61911]: value = "vm-269572" [ 814.168766] env[61911]: _type = "VirtualMachine" [ 814.168766] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 814.169341] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the lease: (returnval){ [ 814.169341] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5266358b-ce95-d68b-2d99-e313d43515c3" [ 814.169341] env[61911]: _type = "HttpNfcLease" [ 814.169341] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 814.169341] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82335efc-8491-463d-84eb-6f6211619845 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.180290] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 814.180290] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527a066f-12c6-9793-1933-6b8076231365" [ 814.180290] env[61911]: _type = "Task" [ 814.180290] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.181804] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 814.181804] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5266358b-ce95-d68b-2d99-e313d43515c3" [ 814.181804] env[61911]: _type = "HttpNfcLease" [ 814.181804] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 814.189200] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-642d279b-da19-4342-b435-a82eb09525d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.190922] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527a066f-12c6-9793-1933-6b8076231365, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.201126] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641ae6df-c414-47b1-9bbf-3402e35771e9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.229662] env[61911]: DEBUG nova.compute.manager [req-5db89a3a-9ad2-41d7-a70e-986f5f760a74 req-29b2f41f-f29f-4ed5-8f15-7e978e653b8a service nova] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Detach interface failed, port_id=5bfc0a97-4a46-47d5-b55d-8d8a0374ff89, reason: Instance 8a363b99-517c-45ff-9e33-e9ceff3a08b0 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 814.240583] env[61911]: DEBUG oslo_vmware.api [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Task: {'id': task-1250993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143206} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.241817] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.241817] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.241817] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.241817] env[61911]: INFO nova.compute.manager [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Took 1.14 seconds to destroy the instance on the hypervisor. [ 814.241817] env[61911]: DEBUG oslo.service.loopingcall [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.241817] env[61911]: DEBUG nova.compute.manager [-] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 814.241817] env[61911]: DEBUG nova.network.neutron [-] [instance: 951e57ba-da09-426a-b500-e5459d346f64] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 814.335235] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250994, 'name': Rename_Task, 'duration_secs': 0.139866} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.335545] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 814.336384] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac9ae26f-acf8-4c4b-a6df-da736ab61abb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.342287] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 814.342287] env[61911]: value = "task-1250996" [ 814.342287] env[61911]: _type = "Task" [ 814.342287] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.348116] env[61911]: DEBUG nova.network.neutron [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Successfully created port: c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.349958] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.350111] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.351102] env[61911]: DEBUG nova.network.neutron [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.355839] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.357472] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a673b3-defd-48fc-bb6f-a35e9961fd69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.364444] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25d6d78-4c8a-4957-870f-6fb226d75978 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.402954] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde7f936-4699-4fd2-9050-37588d9aced2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.413054] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992255a7-64bc-4f2b-b8dd-df41f12b2d5b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.427377] env[61911]: DEBUG nova.compute.provider_tree [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.584793] env[61911]: INFO nova.compute.manager [-] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Took 1.28 seconds to deallocate network for instance. [ 814.681128] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 814.681128] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5266358b-ce95-d68b-2d99-e313d43515c3" [ 814.681128] env[61911]: _type = "HttpNfcLease" [ 814.681128] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 814.681669] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 814.681669] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5266358b-ce95-d68b-2d99-e313d43515c3" [ 814.681669] env[61911]: _type = "HttpNfcLease" [ 814.681669] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 814.682642] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ffeb50-0c07-428b-a650-4915f5e58574 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.706020] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527a066f-12c6-9793-1933-6b8076231365, 'name': SearchDatastore_Task, 'duration_secs': 0.009021} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.706020] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fac4ce-9159-b8cd-3ac1-ce58fc6f8125/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 814.706020] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fac4ce-9159-b8cd-3ac1-ce58fc6f8125/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 814.706020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.706020] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 43d18895-202c-4048-9435-b3484ffd4c07/43d18895-202c-4048-9435-b3484ffd4c07.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 814.707315] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bcee3780-653b-40d0-b974-96857ceb9afa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.776038] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 814.776038] env[61911]: value = "task-1250997" [ 814.776038] env[61911]: _type = "Task" [ 814.776038] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.786929] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1250997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.805619] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-00f308d0-4355-47e3-821d-27e4ef986108 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.862210] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250996, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.894451] env[61911]: DEBUG nova.network.neutron [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.951793] env[61911]: ERROR nova.scheduler.client.report [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [req-b94832a1-99c2-4d21-be13-584fe585c5a4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b8a56394-897a-4e67-92ba-db46db6115e9. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b94832a1-99c2-4d21-be13-584fe585c5a4"}]} [ 814.979328] env[61911]: DEBUG nova.scheduler.client.report [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Refreshing inventories for resource provider b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 814.997619] env[61911]: DEBUG nova.scheduler.client.report [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Updating ProviderTree inventory for provider b8a56394-897a-4e67-92ba-db46db6115e9 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 814.997853] env[61911]: DEBUG nova.compute.provider_tree [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 815.011372] env[61911]: DEBUG nova.scheduler.client.report [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Refreshing aggregate associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, aggregates: None {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 815.035330] env[61911]: DEBUG nova.scheduler.client.report [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Refreshing trait associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 815.041535] env[61911]: DEBUG nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 815.072413] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 815.072734] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.073291] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 815.074247] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.074247] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 815.075478] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 815.075478] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 815.075478] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 815.075636] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 815.075833] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 815.076052] env[61911]: DEBUG nova.virt.hardware [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 815.077357] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c29f01c-0da1-416f-8212-849dd3bb4928 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.089428] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.094238] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6427c19-5a66-46c5-8450-0bd6fef24e47 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.103666] env[61911]: DEBUG nova.network.neutron [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Updating instance_info_cache with network_info: [{"id": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "address": "fa:16:3e:59:fa:2e", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d69ab4c-3b", "ovs_interfaceid": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.288452] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1250997, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504225} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.288730] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 43d18895-202c-4048-9435-b3484ffd4c07/43d18895-202c-4048-9435-b3484ffd4c07.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 815.289063] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.289420] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-801bfb71-1646-4ebf-b4bd-c86d3e341e2b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.302081] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 815.302081] env[61911]: value = "task-1250998" [ 815.302081] env[61911]: _type = "Task" [ 815.302081] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.311323] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1250998, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.356047] env[61911]: DEBUG oslo_vmware.api [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1250996, 'name': PowerOnVM_Task, 'duration_secs': 0.674148} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.359166] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.359531] env[61911]: INFO nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Took 7.65 seconds to spawn the instance on the hypervisor. [ 815.359782] env[61911]: DEBUG nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 815.361680] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93a0c27-2e92-405a-9e64-60d14cc04dd4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.387590] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6996b4f-99c4-4237-899e-b2d22f908d02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.396941] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b1a587-eedd-4eb8-bbb8-f61f50a8ef32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.434326] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdbd6db-aa27-431b-b7c1-2ceacf113f0f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.443941] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c537272-dee9-48f7-a34d-c7f2cfd5467b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.459767] env[61911]: DEBUG nova.compute.provider_tree [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 815.607248] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.607830] env[61911]: DEBUG nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Instance network_info: |[{"id": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "address": "fa:16:3e:59:fa:2e", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d69ab4c-3b", "ovs_interfaceid": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 815.608166] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:fa:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8d69ab4c-3b99-412f-8985-3d8a9210f776', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.616882] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating folder: Project (813b3181120245e594809d096a621675). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.617502] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d38aecac-2e0f-46c9-a9f7-edad912e84d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.628779] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created folder: Project (813b3181120245e594809d096a621675) in parent group-v269521. [ 815.628962] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating folder: Instances. Parent ref: group-v269573. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.629222] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d355f69c-b3e5-4bf5-99d7-8281e5b33fba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.638648] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created folder: Instances in parent group-v269573. [ 815.638958] env[61911]: DEBUG oslo.service.loopingcall [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.639224] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.639451] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3b47f77-5474-46e5-b41f-cbe8538a198b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.659055] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.659055] env[61911]: value = "task-1251001" [ 815.659055] env[61911]: _type = "Task" [ 815.659055] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.666779] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251001, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.712612] env[61911]: DEBUG nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Received event network-changed-8d69ab4c-3b99-412f-8985-3d8a9210f776 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 815.712824] env[61911]: DEBUG nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Refreshing instance network info cache due to event network-changed-8d69ab4c-3b99-412f-8985-3d8a9210f776. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 815.713065] env[61911]: DEBUG oslo_concurrency.lockutils [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] Acquiring lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.713220] env[61911]: DEBUG oslo_concurrency.lockutils [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] Acquired lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.713387] env[61911]: DEBUG nova.network.neutron [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Refreshing network info cache for port 8d69ab4c-3b99-412f-8985-3d8a9210f776 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 815.813789] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1250998, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108655} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.814383] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.815477] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8014b33-02df-468e-883a-5081298a3706 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.840850] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 43d18895-202c-4048-9435-b3484ffd4c07/43d18895-202c-4048-9435-b3484ffd4c07.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.840850] env[61911]: DEBUG nova.network.neutron [-] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.842855] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50b039e5-0163-4455-8501-a1d3fcf99919 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.860566] env[61911]: INFO nova.compute.manager [-] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Took 1.62 seconds to deallocate network for instance. [ 815.869152] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 815.869152] env[61911]: value = "task-1251002" [ 815.869152] env[61911]: _type = "Task" [ 815.869152] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.883821] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251002, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.886153] env[61911]: INFO nova.compute.manager [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Took 31.10 seconds to build instance. [ 815.995146] env[61911]: DEBUG nova.scheduler.client.report [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Updated inventory for provider b8a56394-897a-4e67-92ba-db46db6115e9 with generation 75 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 815.995888] env[61911]: DEBUG nova.compute.provider_tree [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Updating resource provider b8a56394-897a-4e67-92ba-db46db6115e9 generation from 75 to 76 during operation: update_inventory {{(pid=61911) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 815.995888] env[61911]: DEBUG nova.compute.provider_tree [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 816.171500] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251001, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.368044] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.379522] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251002, 'name': ReconfigVM_Task, 'duration_secs': 0.299279} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.380028] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 43d18895-202c-4048-9435-b3484ffd4c07/43d18895-202c-4048-9435-b3484ffd4c07.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.380691] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bf71131-217d-48ed-9962-3d9e1715e08e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.387581] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 816.387581] env[61911]: value = "task-1251003" [ 816.387581] env[61911]: _type = "Task" [ 816.387581] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.387788] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c8dc46ab-f9c0-42f7-b558-df32eb78272d tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "28da8a83-a004-4baf-a744-bcc260032afd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.616s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.398364] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251003, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.419353] env[61911]: DEBUG nova.compute.manager [req-30595991-2ace-4fbf-9d43-0d1b22ef2c2f req-965d878e-051d-4cf8-858d-4a7ac4ef8f51 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-vif-plugged-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 816.419666] env[61911]: DEBUG oslo_concurrency.lockutils [req-30595991-2ace-4fbf-9d43-0d1b22ef2c2f req-965d878e-051d-4cf8-858d-4a7ac4ef8f51 service nova] Acquiring lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.420012] env[61911]: DEBUG oslo_concurrency.lockutils [req-30595991-2ace-4fbf-9d43-0d1b22ef2c2f req-965d878e-051d-4cf8-858d-4a7ac4ef8f51 service nova] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.420299] env[61911]: DEBUG oslo_concurrency.lockutils [req-30595991-2ace-4fbf-9d43-0d1b22ef2c2f req-965d878e-051d-4cf8-858d-4a7ac4ef8f51 service nova] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.420606] env[61911]: DEBUG nova.compute.manager [req-30595991-2ace-4fbf-9d43-0d1b22ef2c2f req-965d878e-051d-4cf8-858d-4a7ac4ef8f51 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] No waiting events found dispatching network-vif-plugged-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 816.420951] env[61911]: WARNING nova.compute.manager [req-30595991-2ace-4fbf-9d43-0d1b22ef2c2f req-965d878e-051d-4cf8-858d-4a7ac4ef8f51 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received unexpected event network-vif-plugged-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 for instance with vm_state building and task_state spawning. [ 816.446688] env[61911]: DEBUG nova.network.neutron [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Successfully updated port: c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.502058] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.482s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.502058] env[61911]: DEBUG nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 816.505842] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.416s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.507605] env[61911]: INFO nova.compute.claims [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.626741] env[61911]: DEBUG nova.network.neutron [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Updated VIF entry in instance network info cache for port 8d69ab4c-3b99-412f-8985-3d8a9210f776. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 816.627453] env[61911]: DEBUG nova.network.neutron [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Updating instance_info_cache with network_info: [{"id": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "address": "fa:16:3e:59:fa:2e", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d69ab4c-3b", "ovs_interfaceid": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.671025] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251001, 'name': CreateVM_Task, 'duration_secs': 0.549536} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.671266] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.672018] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.672225] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.672584] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.672871] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ac24e54-d2ac-4140-b067-a79e532eacbd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.677639] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 816.677639] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52464ee8-9d9a-1ea1-2600-47b62a4f1a1b" [ 816.677639] env[61911]: _type = "Task" [ 816.677639] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.686212] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52464ee8-9d9a-1ea1-2600-47b62a4f1a1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.898755] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251003, 'name': Rename_Task, 'duration_secs': 0.15752} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.899247] env[61911]: DEBUG nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 816.901919] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.903045] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5389cca1-ebc8-42cc-b434-c8c8fa8efbc9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.909698] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 816.909698] env[61911]: value = "task-1251004" [ 816.909698] env[61911]: _type = "Task" [ 816.909698] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.918427] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251004, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.949397] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.949397] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.949397] env[61911]: DEBUG nova.network.neutron [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.012325] env[61911]: DEBUG nova.compute.utils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.013868] env[61911]: DEBUG nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 817.014103] env[61911]: DEBUG nova.network.neutron [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.061593] env[61911]: DEBUG nova.policy [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e6898e0a82a4c18b9f5c6451c19fd07', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9252f5de9c948c8bfd456e742dc1667', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.130839] env[61911]: DEBUG oslo_concurrency.lockutils [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] Releasing lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.131146] env[61911]: DEBUG nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received event network-vif-deleted-fbb4bd3c-3d98-41d5-982a-a668a001e908 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 817.131337] env[61911]: INFO nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Neutron deleted interface fbb4bd3c-3d98-41d5-982a-a668a001e908; detaching it from the instance and deleting it from the info cache [ 817.131613] env[61911]: DEBUG nova.network.neutron [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Updating instance_info_cache with network_info: [{"id": "11106334-7569-4e99-b691-e671ddf7fdaf", "address": "fa:16:3e:93:b3:2e", "network": {"id": "338d9151-b857-4a8a-8d82-6d742771757d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2119394704", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.29", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dbcd0cd6f3cf4c20ba385b874ba34fa1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b8af79a-31d5-4d78-93d7-3919aa1d9186", "external-id": "nsx-vlan-transportzone-324", "segmentation_id": 324, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11106334-75", "ovs_interfaceid": "11106334-7569-4e99-b691-e671ddf7fdaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.188653] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52464ee8-9d9a-1ea1-2600-47b62a4f1a1b, 'name': SearchDatastore_Task, 'duration_secs': 0.009012} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.188986] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.189247] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.189544] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.189628] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.193023] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.193023] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-908ec37c-023c-434d-b3cc-6d22ba0cdddf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.199604] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.199995] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 817.200834] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5871180d-938d-4d27-91f7-effd5c55698c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.206623] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 817.206623] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5263d6f3-5ca9-4588-7384-b0a70d7dbc50" [ 817.206623] env[61911]: _type = "Task" [ 817.206623] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.214769] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5263d6f3-5ca9-4588-7384-b0a70d7dbc50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.432290] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251004, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.432290] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.511213] env[61911]: DEBUG nova.network.neutron [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.519240] env[61911]: DEBUG nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 817.568725] env[61911]: DEBUG nova.network.neutron [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Successfully created port: 8835f0fb-e13b-4847-b5d2-15a1e49d7785 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.635281] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9767fcca-291a-4fdf-ad3d-a91a5092ef3d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.644724] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c07bb95-035b-4a82-85cb-9480b6b2776f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.679032] env[61911]: DEBUG nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Detach interface failed, port_id=fbb4bd3c-3d98-41d5-982a-a668a001e908, reason: Instance 951e57ba-da09-426a-b500-e5459d346f64 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 817.679268] env[61911]: DEBUG nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Received event network-vif-deleted-11106334-7569-4e99-b691-e671ddf7fdaf {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 817.679462] env[61911]: INFO nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Neutron deleted interface 11106334-7569-4e99-b691-e671ddf7fdaf; detaching it from the instance and deleting it from the info cache [ 817.679659] env[61911]: DEBUG nova.network.neutron [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.724025] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5263d6f3-5ca9-4588-7384-b0a70d7dbc50, 'name': SearchDatastore_Task, 'duration_secs': 0.009289} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.724025] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6529fb0-61a5-4498-ab71-528d6add250b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.728637] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 817.728637] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5235db3f-d2ef-d37e-df3f-be20f0091bc7" [ 817.728637] env[61911]: _type = "Task" [ 817.728637] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.743148] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5235db3f-d2ef-d37e-df3f-be20f0091bc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.745429] env[61911]: DEBUG nova.compute.manager [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 817.745626] env[61911]: DEBUG nova.compute.manager [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing instance network info cache due to event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 817.745825] env[61911]: DEBUG oslo_concurrency.lockutils [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.764168] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "28da8a83-a004-4baf-a744-bcc260032afd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.764436] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "28da8a83-a004-4baf-a744-bcc260032afd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.764665] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "28da8a83-a004-4baf-a744-bcc260032afd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.764892] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "28da8a83-a004-4baf-a744-bcc260032afd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.765096] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "28da8a83-a004-4baf-a744-bcc260032afd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.768067] env[61911]: INFO nova.compute.manager [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Terminating instance [ 817.800936] env[61911]: DEBUG nova.network.neutron [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.881356] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb9f514-a49d-40b1-9548-ace07907ec37 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.890175] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a8da55-bb9d-4a26-a5f4-5955352efc6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.929132] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecced36-43f9-4405-a52c-d70022ab4575 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.937991] env[61911]: DEBUG oslo_vmware.api [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251004, 'name': PowerOnVM_Task, 'duration_secs': 0.517198} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.939869] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 817.940165] env[61911]: INFO nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Took 7.84 seconds to spawn the instance on the hypervisor. [ 817.940494] env[61911]: DEBUG nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 817.941460] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37b75f3-a55b-4396-b5cf-5da7fdad6255 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.945831] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4a6299-62aa-4ee7-835c-15d060f27f5a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.967402] env[61911]: DEBUG nova.compute.provider_tree [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.185352] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f52bfce-d3db-40a4-9dcc-4dc8103db159 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.195299] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d9a34d-2e5b-469f-98e1-4c34a533beb5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.223932] env[61911]: DEBUG nova.compute.manager [req-efe541f2-bb50-4bb9-ab4a-c763f80ad091 req-24607aed-0fec-456c-a736-7e6fa86e548d service nova] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Detach interface failed, port_id=11106334-7569-4e99-b691-e671ddf7fdaf, reason: Instance 951e57ba-da09-426a-b500-e5459d346f64 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 818.246030] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5235db3f-d2ef-d37e-df3f-be20f0091bc7, 'name': SearchDatastore_Task, 'duration_secs': 0.01368} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.246030] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.246430] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 31f0a4f7-a69d-4ec4-966d-93f402163b88/31f0a4f7-a69d-4ec4-966d-93f402163b88.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 818.246529] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f23a46a-4f62-4ac6-801a-13fca284c508 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.253415] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 818.253415] env[61911]: value = "task-1251005" [ 818.253415] env[61911]: _type = "Task" [ 818.253415] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.261663] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.273153] env[61911]: DEBUG nova.compute.manager [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 818.273404] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.274320] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f012d391-2e03-4180-a2d2-76f7e9361652 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.283339] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.283339] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c5357f5-9304-4121-af5e-ddad77e15d01 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.289196] env[61911]: DEBUG oslo_vmware.api [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 818.289196] env[61911]: value = "task-1251006" [ 818.289196] env[61911]: _type = "Task" [ 818.289196] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.297387] env[61911]: DEBUG oslo_vmware.api [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1251006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.304012] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.304342] env[61911]: DEBUG nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Instance network_info: |[{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 818.304647] env[61911]: DEBUG oslo_concurrency.lockutils [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.304854] env[61911]: DEBUG nova.network.neutron [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.306129] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:9f:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0501390-f2c7-4b6d-bd50-5d3a03c0e130', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.313742] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Creating folder: Project (6f40f512384d4fedb75bfc53f3195243). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.314359] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d058482c-6063-48e9-9b2d-5022e795e3f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.326523] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Created folder: Project (6f40f512384d4fedb75bfc53f3195243) in parent group-v269521. [ 818.326523] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Creating folder: Instances. Parent ref: group-v269576. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.326523] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-031752f3-2434-404a-8935-4c413f3c715e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.335746] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Created folder: Instances in parent group-v269576. [ 818.336079] env[61911]: DEBUG oslo.service.loopingcall [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.336335] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.336572] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b0f6792-f37a-4c6a-90d3-287f70bcd833 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.356413] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.356413] env[61911]: value = "task-1251009" [ 818.356413] env[61911]: _type = "Task" [ 818.356413] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.364957] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251009, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.477541] env[61911]: DEBUG nova.scheduler.client.report [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 818.484175] env[61911]: INFO nova.compute.manager [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Took 32.14 seconds to build instance. [ 818.531645] env[61911]: DEBUG nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 818.557582] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 818.557870] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.558072] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 818.558272] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.558428] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 818.558584] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 818.558851] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 818.558965] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 818.559159] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 818.559412] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 818.559641] env[61911]: DEBUG nova.virt.hardware [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 818.560591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4d6543-3af6-4793-bcea-8e48e50804df {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.573445] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf90a769-985a-4e48-8af3-bf86a33b2610 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.766349] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251005, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.802236] env[61911]: DEBUG oslo_vmware.api [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1251006, 'name': PowerOffVM_Task, 'duration_secs': 0.39582} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.802548] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 818.802749] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 818.803050] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80f332ac-9665-4e93-9be8-fff3bc8ff3b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.869479] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251009, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.874621] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 818.874949] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 818.875552] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Deleting the datastore file [datastore2] 28da8a83-a004-4baf-a744-bcc260032afd {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.875552] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e864cda7-4365-4eb1-8828-cb63e1e849e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.882731] env[61911]: DEBUG oslo_vmware.api [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for the task: (returnval){ [ 818.882731] env[61911]: value = "task-1251011" [ 818.882731] env[61911]: _type = "Task" [ 818.882731] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.891053] env[61911]: DEBUG oslo_vmware.api [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1251011, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.984824] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.984824] env[61911]: DEBUG nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 818.987586] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.927s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.989099] env[61911]: INFO nova.compute.claims [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.996095] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cddede84-6a69-4ceb-abb5-aeea39cf2023 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.851s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.073311] env[61911]: DEBUG nova.network.neutron [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updated VIF entry in instance network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.073685] env[61911]: DEBUG nova.network.neutron [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.264836] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614401} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.265184] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 31f0a4f7-a69d-4ec4-966d-93f402163b88/31f0a4f7-a69d-4ec4-966d-93f402163b88.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 819.265473] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.265766] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6e5e849-0788-4ecf-a61b-22479c460df1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.272468] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 819.272468] env[61911]: value = "task-1251012" [ 819.272468] env[61911]: _type = "Task" [ 819.272468] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.281441] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251012, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.366637] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251009, 'name': CreateVM_Task, 'duration_secs': 0.609168} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.368067] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.368797] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.368979] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.369334] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.369827] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eac84693-99e6-46c9-ae4b-07e31dc3f12a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.375901] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 819.375901] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad1d6a-5e20-b7e7-3f80-ebfe30df1b83" [ 819.375901] env[61911]: _type = "Task" [ 819.375901] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.384674] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad1d6a-5e20-b7e7-3f80-ebfe30df1b83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.393696] env[61911]: DEBUG oslo_vmware.api [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1251011, 'name': DeleteDatastoreFile_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.503418] env[61911]: DEBUG nova.compute.utils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.503418] env[61911]: DEBUG nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 819.506533] env[61911]: DEBUG nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 819.506533] env[61911]: DEBUG nova.network.neutron [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.529689] env[61911]: DEBUG nova.network.neutron [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Successfully updated port: 8835f0fb-e13b-4847-b5d2-15a1e49d7785 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.575742] env[61911]: DEBUG oslo_concurrency.lockutils [req-a6c63a34-0a55-4f42-9ac9-015daa230336 req-7fe62f5f-b269-40d3-9c76-69174fce1ab4 service nova] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.579795] env[61911]: DEBUG nova.policy [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2266afee766b4660b9be910cd0c080c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f9c20b59cfc4a8389da83b0e6a87730', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.783298] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251012, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077996} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.783647] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.784487] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e618e24d-1dd5-4db8-bed7-1fca96c3c0a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.809159] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 31f0a4f7-a69d-4ec4-966d-93f402163b88/31f0a4f7-a69d-4ec4-966d-93f402163b88.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.809535] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47184f2c-9b08-4bc7-88a1-01e1eba285b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.831074] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 819.831074] env[61911]: value = "task-1251013" [ 819.831074] env[61911]: _type = "Task" [ 819.831074] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.840870] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251013, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.858662] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "43d18895-202c-4048-9435-b3484ffd4c07" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.859021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.859304] env[61911]: INFO nova.compute.manager [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Rebooting instance [ 819.887872] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad1d6a-5e20-b7e7-3f80-ebfe30df1b83, 'name': SearchDatastore_Task, 'duration_secs': 0.021478} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.892075] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.892254] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.892498] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.892662] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.892984] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.894622] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc556d21-cd1a-4f9e-891f-24e2befe8ad7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.897651] env[61911]: DEBUG nova.compute.manager [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Received event network-vif-plugged-8835f0fb-e13b-4847-b5d2-15a1e49d7785 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 819.897862] env[61911]: DEBUG oslo_concurrency.lockutils [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] Acquiring lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.898326] env[61911]: DEBUG oslo_concurrency.lockutils [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.898326] env[61911]: DEBUG oslo_concurrency.lockutils [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.898451] env[61911]: DEBUG nova.compute.manager [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] No waiting events found dispatching network-vif-plugged-8835f0fb-e13b-4847-b5d2-15a1e49d7785 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 819.898584] env[61911]: WARNING nova.compute.manager [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Received unexpected event network-vif-plugged-8835f0fb-e13b-4847-b5d2-15a1e49d7785 for instance with vm_state building and task_state spawning. [ 819.898788] env[61911]: DEBUG nova.compute.manager [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Received event network-changed-8835f0fb-e13b-4847-b5d2-15a1e49d7785 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 819.898898] env[61911]: DEBUG nova.compute.manager [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Refreshing instance network info cache due to event network-changed-8835f0fb-e13b-4847-b5d2-15a1e49d7785. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 819.899097] env[61911]: DEBUG oslo_concurrency.lockutils [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] Acquiring lock "refresh_cache-0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.899240] env[61911]: DEBUG oslo_concurrency.lockutils [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] Acquired lock "refresh_cache-0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.899400] env[61911]: DEBUG nova.network.neutron [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Refreshing network info cache for port 8835f0fb-e13b-4847-b5d2-15a1e49d7785 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.907473] env[61911]: DEBUG oslo_vmware.api [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Task: {'id': task-1251011, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.511771} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.908976] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.909205] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 819.909396] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 819.909578] env[61911]: INFO nova.compute.manager [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Took 1.64 seconds to destroy the instance on the hypervisor. [ 819.909827] env[61911]: DEBUG oslo.service.loopingcall [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.910059] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.910225] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.912783] env[61911]: DEBUG nova.compute.manager [-] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 819.912902] env[61911]: DEBUG nova.network.neutron [-] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 819.914764] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9128ede7-97dc-43d2-8b80-27b528b6eb3a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.921500] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 819.921500] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5263ab60-4120-6981-a547-464acba97df0" [ 819.921500] env[61911]: _type = "Task" [ 819.921500] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.929698] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5263ab60-4120-6981-a547-464acba97df0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.008772] env[61911]: DEBUG nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 820.037506] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "refresh_cache-0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.043783] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.326947] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affc02db-dd81-436e-8e62-2d3d97342aef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.339821] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706110b7-fa78-45ab-a239-25b9cf71d197 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.348418] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251013, 'name': ReconfigVM_Task, 'duration_secs': 0.474602} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.352222] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 31f0a4f7-a69d-4ec4-966d-93f402163b88/31f0a4f7-a69d-4ec4-966d-93f402163b88.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.352222] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3c28570-7041-486c-98b1-862a66f47bda {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.382598] env[61911]: DEBUG nova.network.neutron [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Successfully created port: 15e642fd-30d2-41c2-98f7-0d19e729041e {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.386376] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5c15b4-ad2b-46be-9082-b1ff4d5d4830 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.393220] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 820.393220] env[61911]: value = "task-1251014" [ 820.393220] env[61911]: _type = "Task" [ 820.393220] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.398934] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd2febf-5499-44b0-84bb-8936676227f2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.412351] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251014, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.413948] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.414457] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquired lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.414667] env[61911]: DEBUG nova.network.neutron [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.427670] env[61911]: DEBUG nova.compute.provider_tree [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.440486] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5263ab60-4120-6981-a547-464acba97df0, 'name': SearchDatastore_Task, 'duration_secs': 0.012642} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.441311] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c393e60d-da3b-4a58-84ad-4a91989e59fb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.447073] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 820.447073] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5294bbf7-ddad-3f68-544e-904f2e247704" [ 820.447073] env[61911]: _type = "Task" [ 820.447073] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.455949] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5294bbf7-ddad-3f68-544e-904f2e247704, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.459642] env[61911]: DEBUG nova.network.neutron [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.584177] env[61911]: DEBUG nova.network.neutron [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.901640] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251014, 'name': Rename_Task, 'duration_secs': 0.194344} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.901978] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.902255] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fde8b550-39d1-4580-8531-6c0426a7a037 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.909707] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 820.909707] env[61911]: value = "task-1251015" [ 820.909707] env[61911]: _type = "Task" [ 820.909707] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.925501] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251015, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.932266] env[61911]: DEBUG nova.scheduler.client.report [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 820.963924] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5294bbf7-ddad-3f68-544e-904f2e247704, 'name': SearchDatastore_Task, 'duration_secs': 0.013213} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.965033] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.965033] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] e9833b95-4162-42ba-87a4-d4cc790ac8e5/e9833b95-4162-42ba-87a4-d4cc790ac8e5.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.965033] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7c3a7f4-9924-40c7-b3f7-20582567df85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.973588] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 820.973588] env[61911]: value = "task-1251016" [ 820.973588] env[61911]: _type = "Task" [ 820.973588] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.989324] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251016, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.020661] env[61911]: DEBUG nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 821.052955] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 821.053293] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.053487] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 821.053686] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.053951] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 821.054236] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 821.054537] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 821.054770] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 821.055050] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 821.055328] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 821.055577] env[61911]: DEBUG nova.virt.hardware [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 821.056592] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d829328-c682-4691-bf7b-884f5eb69560 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.066284] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafc9854-8d0a-4c8b-86a2-676439c0bf25 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.087679] env[61911]: DEBUG oslo_concurrency.lockutils [req-090fc4ad-03d2-453c-8e6b-7e1e0a06b2c9 req-920e7439-c4c4-4497-b5b5-7fad42736c80 service nova] Releasing lock "refresh_cache-0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.088104] env[61911]: DEBUG nova.network.neutron [-] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.089370] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquired lock "refresh_cache-0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.089511] env[61911]: DEBUG nova.network.neutron [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.186194] env[61911]: DEBUG nova.network.neutron [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Updating instance_info_cache with network_info: [{"id": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "address": "fa:16:3e:bc:60:07", "network": {"id": "d54e4625-67a5-4f4e-92cd-230f95420309", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2030291444-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a2306e0727146eaa8b8b250bca67966", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7ec41d6-73", "ovs_interfaceid": "d7ec41d6-7384-43bc-86ac-98b776db99e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.422173] env[61911]: DEBUG oslo_vmware.api [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251015, 'name': PowerOnVM_Task, 'duration_secs': 0.495857} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.422548] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.422892] env[61911]: INFO nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Took 8.85 seconds to spawn the instance on the hypervisor. [ 821.423166] env[61911]: DEBUG nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 821.424170] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45eb36ba-effc-4e29-a10e-83f053a42d4d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.441065] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.441065] env[61911]: DEBUG nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 821.442799] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.750s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.444225] env[61911]: INFO nova.compute.claims [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.489845] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251016, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.594145] env[61911]: INFO nova.compute.manager [-] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Took 1.68 seconds to deallocate network for instance. [ 821.624893] env[61911]: DEBUG nova.network.neutron [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.689282] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Releasing lock "refresh_cache-43d18895-202c-4048-9435-b3484ffd4c07" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.766546] env[61911]: DEBUG nova.network.neutron [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Updating instance_info_cache with network_info: [{"id": "8835f0fb-e13b-4847-b5d2-15a1e49d7785", "address": "fa:16:3e:39:8d:cc", "network": {"id": "1e0e9775-fc78-4a6e-bba3-613b20992a61", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-721149936-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9252f5de9c948c8bfd456e742dc1667", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8835f0fb-e1", "ovs_interfaceid": "8835f0fb-e13b-4847-b5d2-15a1e49d7785", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.825542] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fac4ce-9159-b8cd-3ac1-ce58fc6f8125/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 821.826570] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffa60c8-ebd9-4f42-9cda-1e5c8e1c3121 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.834587] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fac4ce-9159-b8cd-3ac1-ce58fc6f8125/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 821.834777] env[61911]: ERROR oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fac4ce-9159-b8cd-3ac1-ce58fc6f8125/disk-0.vmdk due to incomplete transfer. [ 821.835043] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7a84248d-d8b8-44c5-8b32-4f3178fcf70d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.841698] env[61911]: DEBUG oslo_vmware.rw_handles [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fac4ce-9159-b8cd-3ac1-ce58fc6f8125/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 821.841897] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Uploaded image 1dc41913-75eb-4381-b0aa-d373205152d8 to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 821.843937] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 821.844445] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-db6dd45f-e045-45e1-8ea7-d18217a9ab7d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.850028] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 821.850028] env[61911]: value = "task-1251017" [ 821.850028] env[61911]: _type = "Task" [ 821.850028] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.857329] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251017, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.947406] env[61911]: INFO nova.compute.manager [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Took 34.05 seconds to build instance. [ 821.949138] env[61911]: DEBUG nova.compute.utils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.954846] env[61911]: DEBUG nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 821.955048] env[61911]: DEBUG nova.network.neutron [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 821.988326] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251016, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608677} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.988628] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] e9833b95-4162-42ba-87a4-d4cc790ac8e5/e9833b95-4162-42ba-87a4-d4cc790ac8e5.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.988853] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.989146] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-225f56e0-740a-4b58-b269-117beb2e92c8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.997698] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 821.997698] env[61911]: value = "task-1251018" [ 821.997698] env[61911]: _type = "Task" [ 821.997698] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.002273] env[61911]: DEBUG nova.policy [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2266afee766b4660b9be910cd0c080c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f9c20b59cfc4a8389da83b0e6a87730', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.012079] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251018, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.030480] env[61911]: DEBUG nova.compute.manager [req-f128abd5-854c-402f-8d4e-1cb30c97e351 req-91b5e52c-f24b-4e79-93d1-e2825fa2276b service nova] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Received event network-vif-deleted-ea494ffc-7ea2-4dfd-84ed-1e401fe49c16 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 822.101238] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.195276] env[61911]: DEBUG nova.compute.manager [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 822.199815] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262bcc12-9ae6-4205-8462-b8a61cfbeee3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.271682] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Releasing lock "refresh_cache-0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.272211] env[61911]: DEBUG nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Instance network_info: |[{"id": "8835f0fb-e13b-4847-b5d2-15a1e49d7785", "address": "fa:16:3e:39:8d:cc", "network": {"id": "1e0e9775-fc78-4a6e-bba3-613b20992a61", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-721149936-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9252f5de9c948c8bfd456e742dc1667", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8835f0fb-e1", "ovs_interfaceid": "8835f0fb-e13b-4847-b5d2-15a1e49d7785", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 822.272826] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:8d:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92e4d027-e755-417b-8eea-9a8f24b85140', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8835f0fb-e13b-4847-b5d2-15a1e49d7785', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.283434] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Creating folder: Project (c9252f5de9c948c8bfd456e742dc1667). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.283990] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9444d9f7-1193-4dc0-b353-4599211adec4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.300061] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Created folder: Project (c9252f5de9c948c8bfd456e742dc1667) in parent group-v269521. [ 822.300061] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Creating folder: Instances. Parent ref: group-v269579. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.300061] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe9c3db5-f4e7-49d4-b530-6bde65d1b382 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.309237] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Created folder: Instances in parent group-v269579. [ 822.309513] env[61911]: DEBUG oslo.service.loopingcall [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.309741] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.309967] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58b82c76-2848-4845-934f-5deca55e4ae7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.334302] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.334302] env[61911]: value = "task-1251021" [ 822.334302] env[61911]: _type = "Task" [ 822.334302] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.342460] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251021, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.346797] env[61911]: DEBUG nova.network.neutron [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Successfully created port: 122dbb65-8d1c-4797-ba2f-4c534edd83a8 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.358509] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251017, 'name': Destroy_Task, 'duration_secs': 0.331349} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.358828] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Destroyed the VM [ 822.359177] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 822.359416] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-daf24da9-a07e-4697-8708-dbbf3bc63965 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.367269] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 822.367269] env[61911]: value = "task-1251022" [ 822.367269] env[61911]: _type = "Task" [ 822.367269] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.378685] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251022, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.431294] env[61911]: DEBUG nova.network.neutron [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Successfully updated port: 15e642fd-30d2-41c2-98f7-0d19e729041e {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.455475] env[61911]: DEBUG nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 822.458342] env[61911]: DEBUG oslo_concurrency.lockutils [None req-630cc6b5-adeb-4755-a871-a3d877b35540 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.961s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.512857] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251018, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068383} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.512857] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.514762] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e278850-93aa-490e-882e-6e67dbcf5953 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.537236] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] e9833b95-4162-42ba-87a4-d4cc790ac8e5/e9833b95-4162-42ba-87a4-d4cc790ac8e5.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.540703] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d47053e4-5e7f-4b29-901a-898828481f17 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.563964] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 822.563964] env[61911]: value = "task-1251023" [ 822.563964] env[61911]: _type = "Task" [ 822.563964] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.574558] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251023, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.815860] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2795248c-7c87-4452-bf30-c372a769d886 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.823232] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d06de63-0d6c-4496-8faf-83b7144f4ebb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.856131] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8568012-2dfe-4bfa-8345-ccca5defe2c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.867265] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251021, 'name': CreateVM_Task, 'duration_secs': 0.35217} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.869464] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.873269] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.873447] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.873777] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.875060] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4397b5b-7bee-487a-b39c-6526a7a01187 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.879363] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b11ed2ea-3ecc-4ea5-b212-8dbdf80b0726 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.887927] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251022, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.897278] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 822.897278] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52793030-b795-d639-42d3-1595c6a28d7f" [ 822.897278] env[61911]: _type = "Task" [ 822.897278] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.898047] env[61911]: DEBUG nova.compute.provider_tree [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.907919] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52793030-b795-d639-42d3-1595c6a28d7f, 'name': SearchDatastore_Task, 'duration_secs': 0.01394} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.908719] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.908959] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.909211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.909374] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.909561] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.910038] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af291ac4-26f2-4c2d-92b8-660410a468eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.920498] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.920822] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.922346] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0539b8e4-ff23-4321-a3cd-f34f97a01d10 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.928582] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 822.928582] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ba0769-7749-f02f-60d9-bd5455adea3a" [ 822.928582] env[61911]: _type = "Task" [ 822.928582] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.933549] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.933674] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.933826] env[61911]: DEBUG nova.network.neutron [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.940737] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ba0769-7749-f02f-60d9-bd5455adea3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.962741] env[61911]: DEBUG nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 823.074311] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251023, 'name': ReconfigVM_Task, 'duration_secs': 0.293245} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.075525] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Reconfigured VM instance instance-0000003b to attach disk [datastore2] e9833b95-4162-42ba-87a4-d4cc790ac8e5/e9833b95-4162-42ba-87a4-d4cc790ac8e5.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.075525] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c5697b7-bd0d-4bc8-9325-e8af8af1f844 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.081305] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 823.081305] env[61911]: value = "task-1251024" [ 823.081305] env[61911]: _type = "Task" [ 823.081305] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.089163] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251024, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.213568] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3043e3-4507-45b7-b126-23461f88db80 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.221278] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Doing hard reboot of VM {{(pid=61911) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 823.221564] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-4140c3eb-0c93-41d5-8eec-6c4b6a235933 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.227997] env[61911]: DEBUG oslo_vmware.api [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 823.227997] env[61911]: value = "task-1251025" [ 823.227997] env[61911]: _type = "Task" [ 823.227997] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.235724] env[61911]: DEBUG oslo_vmware.api [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251025, 'name': ResetVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.364223] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.367735] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.367735] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.367735] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.367735] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.368606] env[61911]: INFO nova.compute.manager [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Terminating instance [ 823.380357] env[61911]: DEBUG oslo_vmware.api [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251022, 'name': RemoveSnapshot_Task, 'duration_secs': 0.543595} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.381113] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 823.381376] env[61911]: INFO nova.compute.manager [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Took 12.41 seconds to snapshot the instance on the hypervisor. [ 823.403473] env[61911]: DEBUG nova.scheduler.client.report [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 823.443133] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ba0769-7749-f02f-60d9-bd5455adea3a, 'name': SearchDatastore_Task, 'duration_secs': 0.011267} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.444265] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0923dad-c892-4216-8661-172e964aa178 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.450083] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 823.450083] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221a14f-fdb8-4ad7-3d4a-7528f25cc4f5" [ 823.450083] env[61911]: _type = "Task" [ 823.450083] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.460032] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221a14f-fdb8-4ad7-3d4a-7528f25cc4f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.466258] env[61911]: DEBUG nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 823.484348] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.489837] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 823.490096] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.490269] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 823.490492] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.490654] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 823.490807] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 823.491029] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 823.491205] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 823.491382] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 823.491552] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 823.491730] env[61911]: DEBUG nova.virt.hardware [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 823.492674] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb7e6b2-b819-4133-bc8a-41718a3f7aba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.495823] env[61911]: DEBUG nova.network.neutron [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.503255] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150cba75-82b8-4f0a-9ee9-c2cff98acd6e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.590824] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251024, 'name': Rename_Task, 'duration_secs': 0.148428} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.591139] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.591385] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2628b8c-045a-443f-bd58-da88de136842 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.598722] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 823.598722] env[61911]: value = "task-1251026" [ 823.598722] env[61911]: _type = "Task" [ 823.598722] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.610050] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251026, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.643725] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "63d54f69-3d8f-45e9-b317-9d484af954b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.644113] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.727459] env[61911]: DEBUG nova.network.neutron [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Updating instance_info_cache with network_info: [{"id": "15e642fd-30d2-41c2-98f7-0d19e729041e", "address": "fa:16:3e:8e:99:bd", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e642fd-30", "ovs_interfaceid": "15e642fd-30d2-41c2-98f7-0d19e729041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.737905] env[61911]: DEBUG oslo_vmware.api [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251025, 'name': ResetVM_Task, 'duration_secs': 0.097739} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.738209] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Did hard reboot of VM {{(pid=61911) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 823.738419] env[61911]: DEBUG nova.compute.manager [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 823.739730] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d59d6bc-836d-4851-982f-9caffb512cb3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.829183] env[61911]: DEBUG nova.compute.manager [req-be3a7714-d7f6-48ff-9f28-e83fbbcbab84 req-6d64bbbc-a5af-4b27-afa0-b17f5a564b44 service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Received event network-vif-plugged-122dbb65-8d1c-4797-ba2f-4c534edd83a8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 823.829496] env[61911]: DEBUG oslo_concurrency.lockutils [req-be3a7714-d7f6-48ff-9f28-e83fbbcbab84 req-6d64bbbc-a5af-4b27-afa0-b17f5a564b44 service nova] Acquiring lock "c95707ca-7e32-42c0-914a-f0b178f07997-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.829787] env[61911]: DEBUG oslo_concurrency.lockutils [req-be3a7714-d7f6-48ff-9f28-e83fbbcbab84 req-6d64bbbc-a5af-4b27-afa0-b17f5a564b44 service nova] Lock "c95707ca-7e32-42c0-914a-f0b178f07997-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.830056] env[61911]: DEBUG oslo_concurrency.lockutils [req-be3a7714-d7f6-48ff-9f28-e83fbbcbab84 req-6d64bbbc-a5af-4b27-afa0-b17f5a564b44 service nova] Lock "c95707ca-7e32-42c0-914a-f0b178f07997-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.830314] env[61911]: DEBUG nova.compute.manager [req-be3a7714-d7f6-48ff-9f28-e83fbbcbab84 req-6d64bbbc-a5af-4b27-afa0-b17f5a564b44 service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] No waiting events found dispatching network-vif-plugged-122dbb65-8d1c-4797-ba2f-4c534edd83a8 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 823.830570] env[61911]: WARNING nova.compute.manager [req-be3a7714-d7f6-48ff-9f28-e83fbbcbab84 req-6d64bbbc-a5af-4b27-afa0-b17f5a564b44 service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Received unexpected event network-vif-plugged-122dbb65-8d1c-4797-ba2f-4c534edd83a8 for instance with vm_state building and task_state spawning. [ 823.873864] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.874137] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.875903] env[61911]: DEBUG nova.compute.manager [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 823.875985] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.877345] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2501d9c3-4497-45eb-9c67-caaebf3c0d69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.884727] env[61911]: DEBUG nova.compute.manager [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Instance disappeared during snapshot {{(pid=61911) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 823.886638] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 823.887058] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f72e83c-acc9-4f6a-a46a-1a3734ef0658 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.895914] env[61911]: DEBUG nova.compute.manager [None req-6d18a0f8-feda-4b7d-8919-46d824e93d46 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image not found during clean up 1dc41913-75eb-4381-b0aa-d373205152d8 {{(pid=61911) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4600}} [ 823.911059] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.911637] env[61911]: DEBUG nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 823.914319] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.825s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.915780] env[61911]: INFO nova.compute.claims [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.921806] env[61911]: DEBUG nova.network.neutron [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Successfully updated port: 122dbb65-8d1c-4797-ba2f-4c534edd83a8 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.956674] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 823.956971] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 823.957242] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleting the datastore file [datastore1] 35cf6a24-3cc0-40bd-92da-798a56b4e2c9 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.958233] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2185065-a6f0-46b2-8e64-e7307b49a84c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.965925] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221a14f-fdb8-4ad7-3d4a-7528f25cc4f5, 'name': SearchDatastore_Task, 'duration_secs': 0.01484} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.966658] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.966941] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf/0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.967281] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-247fc7ca-a442-4840-aae2-b464ee903edc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.970985] env[61911]: DEBUG oslo_vmware.api [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 823.970985] env[61911]: value = "task-1251028" [ 823.970985] env[61911]: _type = "Task" [ 823.970985] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.976435] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 823.976435] env[61911]: value = "task-1251029" [ 823.976435] env[61911]: _type = "Task" [ 823.976435] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.983558] env[61911]: DEBUG oslo_vmware.api [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251028, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.988838] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251029, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.062036] env[61911]: DEBUG nova.compute.manager [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Received event network-vif-plugged-15e642fd-30d2-41c2-98f7-0d19e729041e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 824.062227] env[61911]: DEBUG oslo_concurrency.lockutils [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] Acquiring lock "bafd58bc-8ae7-49a4-a039-31d328699010-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.062440] env[61911]: DEBUG oslo_concurrency.lockutils [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] Lock "bafd58bc-8ae7-49a4-a039-31d328699010-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.062664] env[61911]: DEBUG oslo_concurrency.lockutils [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] Lock "bafd58bc-8ae7-49a4-a039-31d328699010-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.062885] env[61911]: DEBUG nova.compute.manager [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] No waiting events found dispatching network-vif-plugged-15e642fd-30d2-41c2-98f7-0d19e729041e {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 824.063135] env[61911]: WARNING nova.compute.manager [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Received unexpected event network-vif-plugged-15e642fd-30d2-41c2-98f7-0d19e729041e for instance with vm_state building and task_state spawning. [ 824.063323] env[61911]: DEBUG nova.compute.manager [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Received event network-changed-15e642fd-30d2-41c2-98f7-0d19e729041e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 824.063492] env[61911]: DEBUG nova.compute.manager [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Refreshing instance network info cache due to event network-changed-15e642fd-30d2-41c2-98f7-0d19e729041e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 824.063685] env[61911]: DEBUG oslo_concurrency.lockutils [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] Acquiring lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.108491] env[61911]: DEBUG oslo_vmware.api [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251026, 'name': PowerOnVM_Task, 'duration_secs': 0.47667} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.108693] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.108898] env[61911]: INFO nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Took 9.07 seconds to spawn the instance on the hypervisor. [ 824.109106] env[61911]: DEBUG nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 824.109942] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f1902a-8b4b-4cee-8002-e1a800e8e8f9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.234032] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.234032] env[61911]: DEBUG nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Instance network_info: |[{"id": "15e642fd-30d2-41c2-98f7-0d19e729041e", "address": "fa:16:3e:8e:99:bd", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e642fd-30", "ovs_interfaceid": "15e642fd-30d2-41c2-98f7-0d19e729041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 824.234380] env[61911]: DEBUG oslo_concurrency.lockutils [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] Acquired lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.234596] env[61911]: DEBUG nova.network.neutron [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Refreshing network info cache for port 15e642fd-30d2-41c2-98f7-0d19e729041e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.236104] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:99:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15e642fd-30d2-41c2-98f7-0d19e729041e', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.247710] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating folder: Project (5f9c20b59cfc4a8389da83b0e6a87730). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.251100] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be13e6c0-15d3-4756-b4fa-d5e364b0f083 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.256439] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b13b774-dd26-40fc-9ced-4c0d9c17bbe1 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.397s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.267319] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created folder: Project (5f9c20b59cfc4a8389da83b0e6a87730) in parent group-v269521. [ 824.267659] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating folder: Instances. Parent ref: group-v269582. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.268064] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8dce1d7-1228-48ac-8808-65531e8d4381 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.279369] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created folder: Instances in parent group-v269582. [ 824.279606] env[61911]: DEBUG oslo.service.loopingcall [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.279813] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.280370] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a355daa3-1279-4dd3-8d91-2ba41b52dd12 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.304258] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.304258] env[61911]: value = "task-1251032" [ 824.304258] env[61911]: _type = "Task" [ 824.304258] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.314172] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251032, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.422768] env[61911]: DEBUG nova.compute.utils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.430418] env[61911]: DEBUG nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 824.430418] env[61911]: DEBUG nova.network.neutron [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 824.430418] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.430418] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.430418] env[61911]: DEBUG nova.network.neutron [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.487597] env[61911]: DEBUG oslo_vmware.api [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251028, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239874} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.488709] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.488933] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 824.489236] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 824.489448] env[61911]: INFO nova.compute.manager [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Took 0.61 seconds to destroy the instance on the hypervisor. [ 824.489693] env[61911]: DEBUG oslo.service.loopingcall [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.490496] env[61911]: DEBUG nova.compute.manager [-] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 824.490594] env[61911]: DEBUG nova.network.neutron [-] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.493455] env[61911]: DEBUG nova.policy [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b89d8654369c48f4b39a1a0e8c926bdd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ada275c0bead46f8a9d295952cd930ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.498348] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251029, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518504} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.499124] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf/0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.499356] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.499615] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e40f3079-4e75-4ba1-8443-234c209df40a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.506993] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 824.506993] env[61911]: value = "task-1251033" [ 824.506993] env[61911]: _type = "Task" [ 824.506993] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.519274] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.524020] env[61911]: DEBUG nova.network.neutron [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Updated VIF entry in instance network info cache for port 15e642fd-30d2-41c2-98f7-0d19e729041e. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.524333] env[61911]: DEBUG nova.network.neutron [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Updating instance_info_cache with network_info: [{"id": "15e642fd-30d2-41c2-98f7-0d19e729041e", "address": "fa:16:3e:8e:99:bd", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e642fd-30", "ovs_interfaceid": "15e642fd-30d2-41c2-98f7-0d19e729041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.630019] env[61911]: INFO nova.compute.manager [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Took 34.30 seconds to build instance. [ 824.815489] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251032, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.931628] env[61911]: DEBUG nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 824.972876] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "43d18895-202c-4048-9435-b3484ffd4c07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.973248] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.973910] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "43d18895-202c-4048-9435-b3484ffd4c07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.974065] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.974256] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.976424] env[61911]: DEBUG nova.network.neutron [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.979297] env[61911]: INFO nova.compute.manager [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Terminating instance [ 825.018075] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070766} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.021043] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.024481] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed2be0e-19f4-4781-b2e5-dfd5fd452dbe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.028744] env[61911]: DEBUG oslo_concurrency.lockutils [req-d080a6a0-21d3-4e95-9e66-9e92ba51542b req-25005677-41f0-4d9f-9dda-3557f4ef2637 service nova] Releasing lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.052955] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf/0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.056505] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddd41446-1ee6-4a46-81e5-0c16dd314d71 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.079119] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 825.079119] env[61911]: value = "task-1251034" [ 825.079119] env[61911]: _type = "Task" [ 825.079119] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.089224] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251034, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.130356] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a5f6cada-6e92-42a4-ab67-81f88674d939 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.552s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.165662] env[61911]: DEBUG nova.network.neutron [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Updating instance_info_cache with network_info: [{"id": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "address": "fa:16:3e:86:34:e7", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap122dbb65-8d", "ovs_interfaceid": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.252171] env[61911]: DEBUG nova.network.neutron [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Successfully created port: 28654114-34d7-4037-8226-713299d2e033 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.311171] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaadb71b-39c6-4f3c-bf34-616b92a22069 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.324084] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251032, 'name': CreateVM_Task, 'duration_secs': 0.83818} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.324358] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.325381] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b72ae4-04e8-44c3-b138-d1ea63c7367d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.329122] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.329327] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.329672] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.329962] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c184447f-d39e-4820-b11a-7c8c144bf7c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.335309] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 825.335309] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527bd5f3-f629-1d40-f825-8aea30f374b7" [ 825.335309] env[61911]: _type = "Task" [ 825.335309] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.366638] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153dbaa1-38c4-4915-8c24-a2e5c6e55908 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.373959] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527bd5f3-f629-1d40-f825-8aea30f374b7, 'name': SearchDatastore_Task, 'duration_secs': 0.013459} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.376206] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.376440] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.376704] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.376831] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.377025] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.377334] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0ad28ad-e124-4869-965e-3c254c4d4380 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.380720] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07cd34e1-9109-45dd-98bc-5f9345ee6eb0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.393863] env[61911]: DEBUG nova.compute.provider_tree [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.396095] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.396290] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.397219] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4fe770a-3756-4dad-8827-abadadda2887 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.402252] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 825.402252] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5239a0f7-8ab7-b549-11d2-1a4d0b29a7df" [ 825.402252] env[61911]: _type = "Task" [ 825.402252] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.410312] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5239a0f7-8ab7-b549-11d2-1a4d0b29a7df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.484755] env[61911]: DEBUG nova.compute.manager [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 825.485017] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 825.485936] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cce411-bc44-44e9-a371-07be21fdacb2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.494129] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 825.494390] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-168f6b9e-096a-46e8-b458-b9915d101aa1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.500738] env[61911]: DEBUG oslo_vmware.api [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 825.500738] env[61911]: value = "task-1251035" [ 825.500738] env[61911]: _type = "Task" [ 825.500738] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.509339] env[61911]: DEBUG oslo_vmware.api [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.593821] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.632149] env[61911]: DEBUG nova.network.neutron [-] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.639044] env[61911]: DEBUG nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 825.671525] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.671982] env[61911]: DEBUG nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Instance network_info: |[{"id": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "address": "fa:16:3e:86:34:e7", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap122dbb65-8d", "ovs_interfaceid": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 825.672476] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:34:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '122dbb65-8d1c-4797-ba2f-4c534edd83a8', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.681670] env[61911]: DEBUG oslo.service.loopingcall [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.682812] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.683202] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-231fa9a4-b902-48c1-aca5-a8a23a9be1fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.705192] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.705192] env[61911]: value = "task-1251036" [ 825.705192] env[61911]: _type = "Task" [ 825.705192] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.717754] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251036, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.899255] env[61911]: DEBUG nova.scheduler.client.report [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 825.913720] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5239a0f7-8ab7-b549-11d2-1a4d0b29a7df, 'name': SearchDatastore_Task, 'duration_secs': 0.022506} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.914536] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fec74769-4027-4e51-b7ea-00c37cf1896c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.920240] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 825.920240] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524e8f15-48ad-faed-0967-76043934500f" [ 825.920240] env[61911]: _type = "Task" [ 825.920240] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.928505] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524e8f15-48ad-faed-0967-76043934500f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.944165] env[61911]: DEBUG nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 825.965930] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 825.966207] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.966377] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 825.966566] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.966723] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 825.966872] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 825.967138] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 825.967321] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 825.967527] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 825.967706] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 825.967887] env[61911]: DEBUG nova.virt.hardware [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 825.968740] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010950b8-cb05-4eb6-a404-2426fb4ec36d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.976658] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f931dea5-2c55-49e8-a13a-8ea068eebc89 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.009574] env[61911]: DEBUG oslo_vmware.api [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251035, 'name': PowerOffVM_Task, 'duration_secs': 0.196075} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.009848] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 826.010031] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 826.010283] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46204e51-bb60-4a14-be69-76e5edd9e4d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.024140] env[61911]: DEBUG nova.compute.manager [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Received event network-changed-122dbb65-8d1c-4797-ba2f-4c534edd83a8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 826.024140] env[61911]: DEBUG nova.compute.manager [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Refreshing instance network info cache due to event network-changed-122dbb65-8d1c-4797-ba2f-4c534edd83a8. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 826.024140] env[61911]: DEBUG oslo_concurrency.lockutils [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] Acquiring lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.024140] env[61911]: DEBUG oslo_concurrency.lockutils [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] Acquired lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.024140] env[61911]: DEBUG nova.network.neutron [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Refreshing network info cache for port 122dbb65-8d1c-4797-ba2f-4c534edd83a8 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 826.074542] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 826.074945] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 826.075263] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Deleting the datastore file [datastore2] 43d18895-202c-4048-9435-b3484ffd4c07 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 826.075646] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2330b647-ea82-40b2-ad1b-1b436d7bbd54 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.087268] env[61911]: DEBUG oslo_vmware.api [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for the task: (returnval){ [ 826.087268] env[61911]: value = "task-1251038" [ 826.087268] env[61911]: _type = "Task" [ 826.087268] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.097866] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251034, 'name': ReconfigVM_Task, 'duration_secs': 0.64054} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.098792] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf/0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.099877] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8e74392-160b-4a72-bb4c-1b670aa9e661 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.107405] env[61911]: DEBUG oslo_vmware.api [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251038, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.113091] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 826.113091] env[61911]: value = "task-1251039" [ 826.113091] env[61911]: _type = "Task" [ 826.113091] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.127219] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251039, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.136031] env[61911]: INFO nova.compute.manager [-] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Took 1.65 seconds to deallocate network for instance. [ 826.169376] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.216866] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251036, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.316504] env[61911]: DEBUG nova.compute.manager [req-aceef903-a0a5-4c62-aa83-414548b25ea4 req-88caae49-da46-4d79-937a-92c54a34ad32 service nova] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Received event network-vif-deleted-ddbc2d3e-91d8-4651-a95f-3f20c3ced622 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 826.408069] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.408632] env[61911]: DEBUG nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 826.411198] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.771s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.411390] env[61911]: DEBUG nova.objects.instance [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 826.430692] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524e8f15-48ad-faed-0967-76043934500f, 'name': SearchDatastore_Task, 'duration_secs': 0.009175} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.430957] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.431234] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/bafd58bc-8ae7-49a4-a039-31d328699010.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.431494] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b3e575a-9595-4eb6-b18a-7407e6526fc8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.438471] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 826.438471] env[61911]: value = "task-1251040" [ 826.438471] env[61911]: _type = "Task" [ 826.438471] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.448094] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251040, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.597750] env[61911]: DEBUG oslo_vmware.api [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Task: {'id': task-1251038, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158972} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.598064] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.598299] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 826.598511] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 826.598701] env[61911]: INFO nova.compute.manager [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Took 1.11 seconds to destroy the instance on the hypervisor. [ 826.598960] env[61911]: DEBUG oslo.service.loopingcall [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.599193] env[61911]: DEBUG nova.compute.manager [-] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 826.599305] env[61911]: DEBUG nova.network.neutron [-] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 826.627099] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251039, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.642591] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.715503] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251036, 'name': CreateVM_Task, 'duration_secs': 0.599081} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.717969] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.718699] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.718876] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.719235] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.719831] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42a9c0b6-304c-4822-9b4f-5a27d12828d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.724735] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 826.724735] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524fe11b-79ad-efe7-3acc-3c26a7b32ee7" [ 826.724735] env[61911]: _type = "Task" [ 826.724735] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.732895] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524fe11b-79ad-efe7-3acc-3c26a7b32ee7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.763712] env[61911]: DEBUG nova.network.neutron [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Updated VIF entry in instance network info cache for port 122dbb65-8d1c-4797-ba2f-4c534edd83a8. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.764232] env[61911]: DEBUG nova.network.neutron [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Updating instance_info_cache with network_info: [{"id": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "address": "fa:16:3e:86:34:e7", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap122dbb65-8d", "ovs_interfaceid": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.890283] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.890433] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.915318] env[61911]: DEBUG nova.compute.utils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.916832] env[61911]: DEBUG nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Not allocating networking since 'none' was specified. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 826.948963] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251040, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.126542] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251039, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.237120] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524fe11b-79ad-efe7-3acc-3c26a7b32ee7, 'name': SearchDatastore_Task, 'duration_secs': 0.05233} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.237434] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.238166] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.238166] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.238166] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.238320] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.238508] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-129a85db-abf3-4b18-93b7-452339c49853 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.247440] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.247634] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 827.248364] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de10ea25-388b-45ac-be86-6dda25d5dd66 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.253580] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 827.253580] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5262a19c-eb7f-25ad-3135-50267e4210b4" [ 827.253580] env[61911]: _type = "Task" [ 827.253580] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.261772] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5262a19c-eb7f-25ad-3135-50267e4210b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.270443] env[61911]: DEBUG oslo_concurrency.lockutils [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] Releasing lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.270697] env[61911]: DEBUG nova.compute.manager [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 827.270873] env[61911]: DEBUG nova.compute.manager [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing instance network info cache due to event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 827.271112] env[61911]: DEBUG oslo_concurrency.lockutils [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.271270] env[61911]: DEBUG oslo_concurrency.lockutils [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.271441] env[61911]: DEBUG nova.network.neutron [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.398252] env[61911]: DEBUG nova.network.neutron [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Successfully updated port: 28654114-34d7-4037-8226-713299d2e033 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 827.421959] env[61911]: DEBUG nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 827.427877] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f769c170-d9c7-4a0e-b1c7-005fb46bfc36 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.427877] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.817s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.427980] env[61911]: INFO nova.compute.claims [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.449447] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251040, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534425} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.449573] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/bafd58bc-8ae7-49a4-a039-31d328699010.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.449803] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.450068] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7132d2d4-2dc6-4c08-89ef-2a2323562306 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.456270] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 827.456270] env[61911]: value = "task-1251041" [ 827.456270] env[61911]: _type = "Task" [ 827.456270] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.464418] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251041, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.625208] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251039, 'name': Rename_Task, 'duration_secs': 1.186696} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.625518] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 827.625764] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9d7c82a-17f3-48f0-bf3c-668a12f8ec7e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.631929] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 827.631929] env[61911]: value = "task-1251042" [ 827.631929] env[61911]: _type = "Task" [ 827.631929] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.639221] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.763878] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5262a19c-eb7f-25ad-3135-50267e4210b4, 'name': SearchDatastore_Task, 'duration_secs': 0.013811} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.764678] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b75e9cd-a25e-4a57-85bd-d6d22e00db37 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.769678] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 827.769678] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de2767-3f53-472f-4181-d5bfca52bbc6" [ 827.769678] env[61911]: _type = "Task" [ 827.769678] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.778512] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de2767-3f53-472f-4181-d5bfca52bbc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.857340] env[61911]: DEBUG nova.network.neutron [-] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.900456] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.900650] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.900897] env[61911]: DEBUG nova.network.neutron [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.969048] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251041, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062613} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.969392] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.970125] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79b6441-24d6-4049-94cd-7feddcb5b3a7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.994185] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/bafd58bc-8ae7-49a4-a039-31d328699010.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.994454] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8a7ea23-6783-4526-a7b5-f96e9c9028e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.008751] env[61911]: DEBUG nova.network.neutron [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updated VIF entry in instance network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 828.009166] env[61911]: DEBUG nova.network.neutron [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.016139] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 828.016139] env[61911]: value = "task-1251043" [ 828.016139] env[61911]: _type = "Task" [ 828.016139] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.024560] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.142232] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251042, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.280263] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de2767-3f53-472f-4181-d5bfca52bbc6, 'name': SearchDatastore_Task, 'duration_secs': 0.010851} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.280540] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.280799] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/c95707ca-7e32-42c0-914a-f0b178f07997.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 828.281070] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1451065-4791-4687-9054-47282757decd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.288326] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 828.288326] env[61911]: value = "task-1251044" [ 828.288326] env[61911]: _type = "Task" [ 828.288326] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.298314] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.348735] env[61911]: DEBUG nova.compute.manager [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Received event network-vif-plugged-28654114-34d7-4037-8226-713299d2e033 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 828.349009] env[61911]: DEBUG oslo_concurrency.lockutils [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] Acquiring lock "83616524-2530-4398-9019-148e0658720d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.349296] env[61911]: DEBUG oslo_concurrency.lockutils [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] Lock "83616524-2530-4398-9019-148e0658720d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.349516] env[61911]: DEBUG oslo_concurrency.lockutils [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] Lock "83616524-2530-4398-9019-148e0658720d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.349723] env[61911]: DEBUG nova.compute.manager [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] No waiting events found dispatching network-vif-plugged-28654114-34d7-4037-8226-713299d2e033 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 828.349913] env[61911]: WARNING nova.compute.manager [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Received unexpected event network-vif-plugged-28654114-34d7-4037-8226-713299d2e033 for instance with vm_state building and task_state spawning. [ 828.350156] env[61911]: DEBUG nova.compute.manager [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Received event network-vif-deleted-d7ec41d6-7384-43bc-86ac-98b776db99e8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 828.350286] env[61911]: DEBUG nova.compute.manager [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Received event network-changed-28654114-34d7-4037-8226-713299d2e033 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 828.350458] env[61911]: DEBUG nova.compute.manager [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Refreshing instance network info cache due to event network-changed-28654114-34d7-4037-8226-713299d2e033. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 828.350669] env[61911]: DEBUG oslo_concurrency.lockutils [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] Acquiring lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.360285] env[61911]: INFO nova.compute.manager [-] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Took 1.76 seconds to deallocate network for instance. [ 828.433111] env[61911]: DEBUG nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 828.460201] env[61911]: DEBUG nova.network.neutron [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.469851] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 828.474308] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.474589] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 828.474867] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.475188] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 828.475293] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 828.475538] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 828.475772] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 828.476267] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 828.476267] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 828.476470] env[61911]: DEBUG nova.virt.hardware [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 828.479795] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032e58e0-b76e-4df7-84fe-9e6086a1e5fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.502195] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0794f83f-5e6e-49da-b2b4-745ad612bd36 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.513145] env[61911]: DEBUG oslo_concurrency.lockutils [req-dacbd048-0ecc-4895-a79d-9bf2ebb64473 req-9c66878e-fb9b-42f2-853c-dbc8ecee788e service nova] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.528224] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.535581] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Creating folder: Project (04112ee5b1d84af59f58d12c4d1b8257). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.545471] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b55aa53-b739-4b1a-94f5-5cd614edff67 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.556366] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251043, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.558178] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Created folder: Project (04112ee5b1d84af59f58d12c4d1b8257) in parent group-v269521. [ 828.558414] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Creating folder: Instances. Parent ref: group-v269586. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.561996] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98494da1-c845-4ba8-910f-55c8cbd67f82 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.572990] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Created folder: Instances in parent group-v269586. [ 828.573392] env[61911]: DEBUG oslo.service.loopingcall [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.573616] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.573923] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f9e655b-0c15-494a-8516-eac3483d0caa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.601854] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.601854] env[61911]: value = "task-1251047" [ 828.601854] env[61911]: _type = "Task" [ 828.601854] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.612070] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251047, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.649183] env[61911]: DEBUG oslo_vmware.api [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251042, 'name': PowerOnVM_Task, 'duration_secs': 1.012002} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.649599] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.649994] env[61911]: INFO nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Took 10.12 seconds to spawn the instance on the hypervisor. [ 828.650253] env[61911]: DEBUG nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 828.651491] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5912dad7-9c29-439e-8e94-25e6de62367f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.736035] env[61911]: DEBUG nova.network.neutron [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Updating instance_info_cache with network_info: [{"id": "28654114-34d7-4037-8226-713299d2e033", "address": "fa:16:3e:42:92:99", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28654114-34", "ovs_interfaceid": "28654114-34d7-4037-8226-713299d2e033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.799137] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476475} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.801630] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/c95707ca-7e32-42c0-914a-f0b178f07997.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.801877] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.802333] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbce5eec-9929-46bc-8d67-9056e7f587c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.808500] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 828.808500] env[61911]: value = "task-1251048" [ 828.808500] env[61911]: _type = "Task" [ 828.808500] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.815539] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251048, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.856762] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0634675-3298-4a3b-b7c2-0b752737bddc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.863221] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861b6480-6c74-4196-90fe-6d3bd809445f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.867375] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.894646] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708ca387-f269-4578-8155-ffb2ee1ddce1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.902390] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90ba4c6-b9a3-4515-957c-72f23a5ef108 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.915815] env[61911]: DEBUG nova.compute.provider_tree [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.038749] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251043, 'name': ReconfigVM_Task, 'duration_secs': 0.640681} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.039111] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Reconfigured VM instance instance-0000003d to attach disk [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/bafd58bc-8ae7-49a4-a039-31d328699010.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.039672] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0540328a-5741-4310-bcb3-48935f973530 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.045930] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 829.045930] env[61911]: value = "task-1251049" [ 829.045930] env[61911]: _type = "Task" [ 829.045930] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.053471] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251049, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.113075] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251047, 'name': CreateVM_Task, 'duration_secs': 0.360349} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.113075] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.113300] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.113390] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.113755] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.114124] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2131138a-3aa7-4f19-bf6c-7067ab954394 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.119188] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 829.119188] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524a2fb7-92f4-3ec5-8a62-9f461adc9c2e" [ 829.119188] env[61911]: _type = "Task" [ 829.119188] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.127570] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524a2fb7-92f4-3ec5-8a62-9f461adc9c2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.177689] env[61911]: INFO nova.compute.manager [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Took 37.07 seconds to build instance. [ 829.242841] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.243800] env[61911]: DEBUG nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Instance network_info: |[{"id": "28654114-34d7-4037-8226-713299d2e033", "address": "fa:16:3e:42:92:99", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28654114-34", "ovs_interfaceid": "28654114-34d7-4037-8226-713299d2e033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 829.243800] env[61911]: DEBUG oslo_concurrency.lockutils [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] Acquired lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.243800] env[61911]: DEBUG nova.network.neutron [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Refreshing network info cache for port 28654114-34d7-4037-8226-713299d2e033 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 829.244965] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:92:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '098df9b7-d759-47f7-b756-334848cb423b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28654114-34d7-4037-8226-713299d2e033', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.253097] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Creating folder: Project (ada275c0bead46f8a9d295952cd930ec). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.256170] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a733477b-4c07-470e-a856-d5eb7b5753d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.267817] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Created folder: Project (ada275c0bead46f8a9d295952cd930ec) in parent group-v269521. [ 829.268123] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Creating folder: Instances. Parent ref: group-v269589. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.268407] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-641e4939-9c00-48c9-94c6-44ae9bc607bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.278283] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Created folder: Instances in parent group-v269589. [ 829.278568] env[61911]: DEBUG oslo.service.loopingcall [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.278775] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83616524-2530-4398-9019-148e0658720d] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.278988] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f466ac5-ab81-4526-8f14-74cd14afbf5a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.298372] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.298372] env[61911]: value = "task-1251052" [ 829.298372] env[61911]: _type = "Task" [ 829.298372] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.306321] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251052, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.316561] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251048, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095519} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.316844] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.319713] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5598cf00-ced9-481c-8fd8-a7bce4db8ebd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.342427] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/c95707ca-7e32-42c0-914a-f0b178f07997.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.342717] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8815dfb2-bd14-4fd1-a060-83beb6a804b5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.362382] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 829.362382] env[61911]: value = "task-1251053" [ 829.362382] env[61911]: _type = "Task" [ 829.362382] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.371203] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251053, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.420609] env[61911]: DEBUG nova.scheduler.client.report [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 829.558016] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251049, 'name': Rename_Task, 'duration_secs': 0.133487} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.558447] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.558822] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-352b54f1-189f-48b3-aea9-7af63ab8cfb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.566457] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 829.566457] env[61911]: value = "task-1251054" [ 829.566457] env[61911]: _type = "Task" [ 829.566457] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.576960] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.609213] env[61911]: DEBUG nova.network.neutron [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Updated VIF entry in instance network info cache for port 28654114-34d7-4037-8226-713299d2e033. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.609825] env[61911]: DEBUG nova.network.neutron [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Updating instance_info_cache with network_info: [{"id": "28654114-34d7-4037-8226-713299d2e033", "address": "fa:16:3e:42:92:99", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28654114-34", "ovs_interfaceid": "28654114-34d7-4037-8226-713299d2e033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.632061] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524a2fb7-92f4-3ec5-8a62-9f461adc9c2e, 'name': SearchDatastore_Task, 'duration_secs': 0.009827} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.632601] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.632738] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.632903] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.633073] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.633292] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.633575] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9623998d-6d3a-436b-ad35-f78a6c25131a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.643281] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.643466] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.644466] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc0835b7-5b4c-4fd4-a0b4-80188c43398e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.649498] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 829.649498] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d519fa-8f54-fe66-ebac-ef2d5d78e4ee" [ 829.649498] env[61911]: _type = "Task" [ 829.649498] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.657452] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d519fa-8f54-fe66-ebac-ef2d5d78e4ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.679690] env[61911]: DEBUG oslo_concurrency.lockutils [None req-01a2a11c-2620-4d85-8344-32b4fc3a204c tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.607s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.808233] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251052, 'name': CreateVM_Task, 'duration_secs': 0.357725} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.808409] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83616524-2530-4398-9019-148e0658720d] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.809113] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.809308] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.809663] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.809944] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1b2923d-d733-4787-9064-71449eefdb7d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.814361] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 829.814361] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524080c3-b58a-4bc2-779f-9737970cd723" [ 829.814361] env[61911]: _type = "Task" [ 829.814361] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.821886] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524080c3-b58a-4bc2-779f-9737970cd723, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.873226] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251053, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.927565] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.928252] env[61911]: DEBUG nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 829.931100] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.842s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.931374] env[61911]: DEBUG nova.objects.instance [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lazy-loading 'resources' on Instance uuid 8a363b99-517c-45ff-9e33-e9ceff3a08b0 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.077472] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251054, 'name': PowerOnVM_Task} progress is 96%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.112769] env[61911]: DEBUG oslo_concurrency.lockutils [req-e75b91cb-7bd8-4c09-9988-70e00e9ee857 req-6d8a0c7d-47a7-43dd-8a1d-a598693cdfa0 service nova] Releasing lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.159705] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d519fa-8f54-fe66-ebac-ef2d5d78e4ee, 'name': SearchDatastore_Task, 'duration_secs': 0.008661} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.160648] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7af34fca-f15a-455c-afbf-7c4651bb35e2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.165906] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 830.165906] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e4aab3-78fe-7d90-4a2c-56c1f4587af0" [ 830.165906] env[61911]: _type = "Task" [ 830.165906] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.173423] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e4aab3-78fe-7d90-4a2c-56c1f4587af0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.181998] env[61911]: DEBUG nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 830.325423] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524080c3-b58a-4bc2-779f-9737970cd723, 'name': SearchDatastore_Task, 'duration_secs': 0.008862} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.325739] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.325978] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.326221] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.373550] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251053, 'name': ReconfigVM_Task, 'duration_secs': 0.570905} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.373985] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Reconfigured VM instance instance-0000003e to attach disk [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/c95707ca-7e32-42c0-914a-f0b178f07997.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.374842] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1634a89-8a9d-4739-8905-1be63590749c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.381362] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 830.381362] env[61911]: value = "task-1251055" [ 830.381362] env[61911]: _type = "Task" [ 830.381362] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.389852] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251055, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.434276] env[61911]: DEBUG nova.compute.utils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.438733] env[61911]: DEBUG nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 830.438937] env[61911]: DEBUG nova.network.neutron [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.492053] env[61911]: DEBUG nova.policy [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8593a447cc994e1eb0d61a548ba98ba8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e38ba711d3e3431b9d258c91ee1ce6f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.582869] env[61911]: DEBUG oslo_vmware.api [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251054, 'name': PowerOnVM_Task, 'duration_secs': 0.544501} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.583247] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.583510] env[61911]: INFO nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Took 9.56 seconds to spawn the instance on the hypervisor. [ 830.583773] env[61911]: DEBUG nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.584958] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418bacd4-97de-489b-81f5-267dbb6025a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.679050] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e4aab3-78fe-7d90-4a2c-56c1f4587af0, 'name': SearchDatastore_Task, 'duration_secs': 0.008744} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.679385] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.679612] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 65d310c6-d438-4c3d-bd44-f3b51123fe93/65d310c6-d438-4c3d-bd44-f3b51123fe93.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.680118] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.680336] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.680569] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-892d1efe-951f-4b2a-b56d-08bf9efa2931 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.685399] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62a6583e-26ef-42b1-ae4c-d2c65265dfe9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.695728] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 830.695728] env[61911]: value = "task-1251056" [ 830.695728] env[61911]: _type = "Task" [ 830.695728] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.698557] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.698752] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.701993] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1541fb23-ce4a-4ead-ac01-d0e705c22cb4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.704926] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.708067] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251056, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.710937] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 830.710937] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522d6675-2cdd-7a77-f81c-db8c1242feb6" [ 830.710937] env[61911]: _type = "Task" [ 830.710937] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.720797] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522d6675-2cdd-7a77-f81c-db8c1242feb6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.776005] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f712b272-947b-4b2e-86e4-b04335fca732 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.786809] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e51d5c-6191-4bea-8768-d3b057871351 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.816360] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918a2707-a7d5-4715-9501-8d5d4fe6bc66 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.823603] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33905cb-e84f-4f4c-a477-4ab143715e8c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.836164] env[61911]: DEBUG nova.compute.provider_tree [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.893390] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251055, 'name': Rename_Task, 'duration_secs': 0.21124} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.893750] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.894039] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-226f21c9-eca1-4069-99a1-b9e58268c98a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.900888] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 830.900888] env[61911]: value = "task-1251057" [ 830.900888] env[61911]: _type = "Task" [ 830.900888] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.908545] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251057, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.939941] env[61911]: DEBUG nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 830.986527] env[61911]: DEBUG nova.network.neutron [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Successfully created port: 2ce1966f-9d77-419b-8508-acc03ab5d271 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.108159] env[61911]: INFO nova.compute.manager [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Took 34.04 seconds to build instance. [ 831.204172] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251056, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.221219] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522d6675-2cdd-7a77-f81c-db8c1242feb6, 'name': SearchDatastore_Task, 'duration_secs': 0.010509} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.222246] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd65b15b-d345-4f57-858f-383dbd3de3dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.228083] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 831.228083] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526cb307-1187-a8e3-dbdc-d3ab24151e1b" [ 831.228083] env[61911]: _type = "Task" [ 831.228083] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.237561] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526cb307-1187-a8e3-dbdc-d3ab24151e1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.339630] env[61911]: DEBUG nova.scheduler.client.report [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 831.414425] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251057, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.473937] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.474135] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.474372] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.474571] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.474752] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.477013] env[61911]: INFO nova.compute.manager [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Terminating instance [ 831.610257] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a4c844a-b519-4acb-9e4b-2cf5f2eed724 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "bafd58bc-8ae7-49a4-a039-31d328699010" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.806s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.703865] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251056, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559559} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.704196] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 65d310c6-d438-4c3d-bd44-f3b51123fe93/65d310c6-d438-4c3d-bd44-f3b51123fe93.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.704423] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.704685] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8bf51ac-6a12-441e-837a-679dbdbb21bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.712024] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 831.712024] env[61911]: value = "task-1251058" [ 831.712024] env[61911]: _type = "Task" [ 831.712024] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.719476] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251058, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.737792] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526cb307-1187-a8e3-dbdc-d3ab24151e1b, 'name': SearchDatastore_Task, 'duration_secs': 0.015722} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.738342] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.738427] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 83616524-2530-4398-9019-148e0658720d/83616524-2530-4398-9019-148e0658720d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.738666] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce729822-e35b-43b7-9790-81fab13d4468 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.744513] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 831.744513] env[61911]: value = "task-1251059" [ 831.744513] env[61911]: _type = "Task" [ 831.744513] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.754307] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251059, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.845367] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.847764] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.480s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.847998] env[61911]: DEBUG nova.objects.instance [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lazy-loading 'resources' on Instance uuid 951e57ba-da09-426a-b500-e5459d346f64 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.871440] env[61911]: INFO nova.scheduler.client.report [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted allocations for instance 8a363b99-517c-45ff-9e33-e9ceff3a08b0 [ 831.912775] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251057, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.948758] env[61911]: DEBUG nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 831.978498] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 831.978866] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.979094] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 831.979342] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.979538] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 831.979702] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 831.980124] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 831.980124] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 831.980280] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 831.980456] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 831.980699] env[61911]: DEBUG nova.virt.hardware [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 831.981959] env[61911]: DEBUG nova.compute.manager [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 831.981959] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.982734] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09535c5-8466-4c96-9230-cba3a978b1fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.986354] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0099a639-fcff-4553-88d6-759c2ed8a317 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.999199] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d824332-be7b-4802-923a-69401a45275a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.003706] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.004501] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0795a87-78e5-456b-865f-9638868508cd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.020034] env[61911]: DEBUG oslo_vmware.api [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 832.020034] env[61911]: value = "task-1251060" [ 832.020034] env[61911]: _type = "Task" [ 832.020034] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.028377] env[61911]: DEBUG oslo_vmware.api [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.113230] env[61911]: DEBUG nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 832.222507] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251058, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071212} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.222853] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.223763] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6198af-02a0-4eb1-9d0f-01b0ff0c09cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.245064] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 65d310c6-d438-4c3d-bd44-f3b51123fe93/65d310c6-d438-4c3d-bd44-f3b51123fe93.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.245713] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aec50691-e851-4a72-99ff-a85beffc0790 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.269059] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251059, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501475} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.270517] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 83616524-2530-4398-9019-148e0658720d/83616524-2530-4398-9019-148e0658720d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.270785] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.271382] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 832.271382] env[61911]: value = "task-1251061" [ 832.271382] env[61911]: _type = "Task" [ 832.271382] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.271382] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b30dc1a1-37ab-4e87-83e0-ea6e4ebb5495 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.282248] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251061, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.283642] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 832.283642] env[61911]: value = "task-1251062" [ 832.283642] env[61911]: _type = "Task" [ 832.283642] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.293431] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.382482] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4def6fe5-10dd-419f-a4cf-1e5207d66a18 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "8a363b99-517c-45ff-9e33-e9ceff3a08b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.774s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.419554] env[61911]: DEBUG oslo_vmware.api [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251057, 'name': PowerOnVM_Task, 'duration_secs': 1.405424} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.423053] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.423386] env[61911]: INFO nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Took 8.96 seconds to spawn the instance on the hypervisor. [ 832.423661] env[61911]: DEBUG nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 832.425010] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c12a398-3e1b-4555-b2cd-ad16cd5c2c69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.531338] env[61911]: DEBUG oslo_vmware.api [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251060, 'name': PowerOffVM_Task, 'duration_secs': 0.323242} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.531672] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.531841] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.532118] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b3fdf98-3ee1-41fa-8e9f-28ec1a1a9e22 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.600749] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.600909] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.601113] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Deleting the datastore file [datastore2] 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.601376] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-967f88bb-00dd-4756-9611-99e9f4bef53c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.607930] env[61911]: DEBUG oslo_vmware.api [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for the task: (returnval){ [ 832.607930] env[61911]: value = "task-1251064" [ 832.607930] env[61911]: _type = "Task" [ 832.607930] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.616152] env[61911]: DEBUG oslo_vmware.api [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.634012] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.640591] env[61911]: DEBUG nova.compute.manager [req-ade0f86e-b14f-4e58-95e6-542ba92e5e07 req-c98b3af3-2513-4022-bc85-371c9d294ea1 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Received event network-vif-plugged-2ce1966f-9d77-419b-8508-acc03ab5d271 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 832.640815] env[61911]: DEBUG oslo_concurrency.lockutils [req-ade0f86e-b14f-4e58-95e6-542ba92e5e07 req-c98b3af3-2513-4022-bc85-371c9d294ea1 service nova] Acquiring lock "613544bc-92b9-4c8f-8a7b-f02205f347ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.641066] env[61911]: DEBUG oslo_concurrency.lockutils [req-ade0f86e-b14f-4e58-95e6-542ba92e5e07 req-c98b3af3-2513-4022-bc85-371c9d294ea1 service nova] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.641276] env[61911]: DEBUG oslo_concurrency.lockutils [req-ade0f86e-b14f-4e58-95e6-542ba92e5e07 req-c98b3af3-2513-4022-bc85-371c9d294ea1 service nova] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.641473] env[61911]: DEBUG nova.compute.manager [req-ade0f86e-b14f-4e58-95e6-542ba92e5e07 req-c98b3af3-2513-4022-bc85-371c9d294ea1 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] No waiting events found dispatching network-vif-plugged-2ce1966f-9d77-419b-8508-acc03ab5d271 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 832.641677] env[61911]: WARNING nova.compute.manager [req-ade0f86e-b14f-4e58-95e6-542ba92e5e07 req-c98b3af3-2513-4022-bc85-371c9d294ea1 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Received unexpected event network-vif-plugged-2ce1966f-9d77-419b-8508-acc03ab5d271 for instance with vm_state building and task_state spawning. [ 832.673233] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb959df-925d-427f-8985-9b081dfb692d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.681543] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b127c1-69a0-43f3-a36a-f0765484ee88 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.713171] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d12829-577b-4aea-bf71-6110ad55f050 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.721213] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83936e8b-6e30-4115-9983-61411db2254a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.734674] env[61911]: DEBUG nova.compute.provider_tree [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.758015] env[61911]: DEBUG nova.network.neutron [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Successfully updated port: 2ce1966f-9d77-419b-8508-acc03ab5d271 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.785009] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251061, 'name': ReconfigVM_Task, 'duration_secs': 0.304124} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.788791] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 65d310c6-d438-4c3d-bd44-f3b51123fe93/65d310c6-d438-4c3d-bd44-f3b51123fe93.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.788997] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d5718f6-7a24-4e1c-8e03-263a79678ccf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.795149] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074277} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.796364] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.796718] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 832.796718] env[61911]: value = "task-1251065" [ 832.796718] env[61911]: _type = "Task" [ 832.796718] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.797408] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4025d99b-6ba7-4ddd-8562-500d1bf4b0e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.821825] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 83616524-2530-4398-9019-148e0658720d/83616524-2530-4398-9019-148e0658720d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.825387] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c4fdb7d-dc4f-4001-86f5-d2a78118a88d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.840190] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251065, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.846479] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 832.846479] env[61911]: value = "task-1251066" [ 832.846479] env[61911]: _type = "Task" [ 832.846479] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.855157] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251066, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.948848] env[61911]: INFO nova.compute.manager [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Took 30.90 seconds to build instance. [ 833.118769] env[61911]: DEBUG oslo_vmware.api [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Task: {'id': task-1251064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131177} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.119258] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.119595] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.119886] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.120138] env[61911]: INFO nova.compute.manager [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 833.120478] env[61911]: DEBUG oslo.service.loopingcall [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.120735] env[61911]: DEBUG nova.compute.manager [-] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 833.120865] env[61911]: DEBUG nova.network.neutron [-] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.237557] env[61911]: DEBUG nova.scheduler.client.report [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 833.260507] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "refresh_cache-613544bc-92b9-4c8f-8a7b-f02205f347ff" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.260507] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "refresh_cache-613544bc-92b9-4c8f-8a7b-f02205f347ff" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.260889] env[61911]: DEBUG nova.network.neutron [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.313023] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251065, 'name': Rename_Task, 'duration_secs': 0.177967} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.313541] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.314340] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b105bd9-5274-4c48-8ff5-bacc7b81c4ed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.322062] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 833.322062] env[61911]: value = "task-1251067" [ 833.322062] env[61911]: _type = "Task" [ 833.322062] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.330917] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.357091] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251066, 'name': ReconfigVM_Task, 'duration_secs': 0.316779} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.361053] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 83616524-2530-4398-9019-148e0658720d/83616524-2530-4398-9019-148e0658720d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.361053] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ce5e0da-8a7e-4302-a8fd-3413ed24cb04 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.365060] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 833.365060] env[61911]: value = "task-1251068" [ 833.365060] env[61911]: _type = "Task" [ 833.365060] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.373934] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251068, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.450486] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7fa08ac1-d902-4ef5-8676-56117c92b493 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "c95707ca-7e32-42c0-914a-f0b178f07997" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.170s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.564151] env[61911]: INFO nova.compute.manager [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Rescuing [ 833.564439] env[61911]: DEBUG oslo_concurrency.lockutils [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.564615] env[61911]: DEBUG oslo_concurrency.lockutils [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.564792] env[61911]: DEBUG nova.network.neutron [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.742298] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.744915] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.314s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.746512] env[61911]: INFO nova.compute.claims [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.771156] env[61911]: INFO nova.scheduler.client.report [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Deleted allocations for instance 951e57ba-da09-426a-b500-e5459d346f64 [ 833.798215] env[61911]: DEBUG nova.network.neutron [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.831356] env[61911]: DEBUG oslo_vmware.api [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251067, 'name': PowerOnVM_Task, 'duration_secs': 0.477975} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.831637] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.831839] env[61911]: INFO nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Took 5.40 seconds to spawn the instance on the hypervisor. [ 833.832036] env[61911]: DEBUG nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 833.832806] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a964a5e6-4557-462d-8766-6357c99892eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.874228] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251068, 'name': Rename_Task, 'duration_secs': 0.16189} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.875205] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.875492] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-340ce84b-db1f-4142-8b70-02e6c33fe6cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.884186] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 833.884186] env[61911]: value = "task-1251069" [ 833.884186] env[61911]: _type = "Task" [ 833.884186] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.891901] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.960627] env[61911]: DEBUG nova.network.neutron [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Updating instance_info_cache with network_info: [{"id": "2ce1966f-9d77-419b-8508-acc03ab5d271", "address": "fa:16:3e:3b:7c:75", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ce1966f-9d", "ovs_interfaceid": "2ce1966f-9d77-419b-8508-acc03ab5d271", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.977585] env[61911]: DEBUG nova.network.neutron [-] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.279250] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c9bbcbdb-2654-4a7f-b758-f4fdd3351101 tempest-ServersTestMultiNic-1507407774 tempest-ServersTestMultiNic-1507407774-project-member] Lock "951e57ba-da09-426a-b500-e5459d346f64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.694s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.305164] env[61911]: DEBUG nova.network.neutron [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Updating instance_info_cache with network_info: [{"id": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "address": "fa:16:3e:86:34:e7", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap122dbb65-8d", "ovs_interfaceid": "122dbb65-8d1c-4797-ba2f-4c534edd83a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.353936] env[61911]: INFO nova.compute.manager [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Took 27.28 seconds to build instance. [ 834.396030] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251069, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.410404] env[61911]: DEBUG oslo_concurrency.lockutils [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "28159e75-9fe9-44c7-b5c9-534902cecbef" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.410404] env[61911]: DEBUG oslo_concurrency.lockutils [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.463460] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "refresh_cache-613544bc-92b9-4c8f-8a7b-f02205f347ff" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.463791] env[61911]: DEBUG nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Instance network_info: |[{"id": "2ce1966f-9d77-419b-8508-acc03ab5d271", "address": "fa:16:3e:3b:7c:75", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ce1966f-9d", "ovs_interfaceid": "2ce1966f-9d77-419b-8508-acc03ab5d271", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 834.464237] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:7c:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ce1966f-9d77-419b-8508-acc03ab5d271', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.472134] env[61911]: DEBUG oslo.service.loopingcall [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.472353] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 834.472624] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a540e707-3d05-45d9-af5c-cc72f4183a12 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.489913] env[61911]: INFO nova.compute.manager [-] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Took 1.37 seconds to deallocate network for instance. [ 834.497199] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.497199] env[61911]: value = "task-1251070" [ 834.497199] env[61911]: _type = "Task" [ 834.497199] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.506353] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251070, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.800750] env[61911]: DEBUG nova.compute.manager [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Received event network-changed-2ce1966f-9d77-419b-8508-acc03ab5d271 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 834.801201] env[61911]: DEBUG nova.compute.manager [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Refreshing instance network info cache due to event network-changed-2ce1966f-9d77-419b-8508-acc03ab5d271. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 834.801526] env[61911]: DEBUG oslo_concurrency.lockutils [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] Acquiring lock "refresh_cache-613544bc-92b9-4c8f-8a7b-f02205f347ff" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.801816] env[61911]: DEBUG oslo_concurrency.lockutils [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] Acquired lock "refresh_cache-613544bc-92b9-4c8f-8a7b-f02205f347ff" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.802561] env[61911]: DEBUG nova.network.neutron [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Refreshing network info cache for port 2ce1966f-9d77-419b-8508-acc03ab5d271 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.810856] env[61911]: DEBUG oslo_concurrency.lockutils [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-c95707ca-7e32-42c0-914a-f0b178f07997" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.851243] env[61911]: DEBUG nova.compute.manager [None req-2827d893-5903-4c2c-9c68-f3f7ff41a261 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 834.851243] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961d4c1b-ba0c-45e9-877d-991e02d8197e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.856440] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5ca7f27a-338c-46cf-b1c9-59f7f76c05f1 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "65d310c6-d438-4c3d-bd44-f3b51123fe93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.093s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.898440] env[61911]: DEBUG oslo_vmware.api [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251069, 'name': PowerOnVM_Task, 'duration_secs': 0.865471} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.898988] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 834.899213] env[61911]: INFO nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Took 8.95 seconds to spawn the instance on the hypervisor. [ 834.899804] env[61911]: DEBUG nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 834.900237] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df5b1a9-eba6-4d11-aab8-24f7d17a2ed9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.913625] env[61911]: DEBUG nova.compute.utils [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.999093] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.010843] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251070, 'name': CreateVM_Task, 'duration_secs': 0.344714} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.010942] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 835.011795] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.012036] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.012393] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.012724] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27a9e1da-2a24-42f6-8c04-a1fe3b0cf0b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.021015] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 835.021015] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e246f6-41b8-8687-8a28-7021a355afb8" [ 835.021015] env[61911]: _type = "Task" [ 835.021015] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.028236] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e246f6-41b8-8687-8a28-7021a355afb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.054183] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "65d310c6-d438-4c3d-bd44-f3b51123fe93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.054447] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "65d310c6-d438-4c3d-bd44-f3b51123fe93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.054658] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "65d310c6-d438-4c3d-bd44-f3b51123fe93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.054983] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "65d310c6-d438-4c3d-bd44-f3b51123fe93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.055104] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "65d310c6-d438-4c3d-bd44-f3b51123fe93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.061493] env[61911]: INFO nova.compute.manager [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Terminating instance [ 835.084827] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab40990-15d9-4fbd-802f-ebc4d2c31082 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.093014] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad42017-05b5-4ee0-8800-1367c5b6f0d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.130217] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b95d822-9f73-424a-b1b5-d92fa82718e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.137168] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19139fc3-988b-4b91-a669-407961dc5ef7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.151478] env[61911]: DEBUG nova.compute.provider_tree [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.363649] env[61911]: INFO nova.compute.manager [None req-2827d893-5903-4c2c-9c68-f3f7ff41a261 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] instance snapshotting [ 835.364229] env[61911]: DEBUG nova.objects.instance [None req-2827d893-5903-4c2c-9c68-f3f7ff41a261 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lazy-loading 'flavor' on Instance uuid 65d310c6-d438-4c3d-bd44-f3b51123fe93 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.419621] env[61911]: DEBUG oslo_concurrency.lockutils [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.423460] env[61911]: INFO nova.compute.manager [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Took 32.75 seconds to build instance. [ 835.530376] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e246f6-41b8-8687-8a28-7021a355afb8, 'name': SearchDatastore_Task, 'duration_secs': 0.013722} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.530695] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.530934] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.531187] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.531337] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.531517] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.531781] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-047c7cc2-2cce-4da8-8069-ee7a95301006 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.543519] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.543713] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.544427] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe18dc1d-b73f-46b5-a9e0-def95f60ee85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.549422] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 835.549422] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528e478e-2ea7-c1b9-ef85-2d88c68a4d9a" [ 835.549422] env[61911]: _type = "Task" [ 835.549422] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.556491] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528e478e-2ea7-c1b9-ef85-2d88c68a4d9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.565536] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "refresh_cache-65d310c6-d438-4c3d-bd44-f3b51123fe93" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.565726] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquired lock "refresh_cache-65d310c6-d438-4c3d-bd44-f3b51123fe93" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.565910] env[61911]: DEBUG nova.network.neutron [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.658895] env[61911]: DEBUG nova.scheduler.client.report [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 835.870303] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6226ce24-4260-4b74-b768-7df0b754dd07 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.890322] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea25ee9-a6ed-43cd-8285-5fcc47959b7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.918801] env[61911]: DEBUG nova.network.neutron [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Updated VIF entry in instance network info cache for port 2ce1966f-9d77-419b-8508-acc03ab5d271. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.919166] env[61911]: DEBUG nova.network.neutron [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Updating instance_info_cache with network_info: [{"id": "2ce1966f-9d77-419b-8508-acc03ab5d271", "address": "fa:16:3e:3b:7c:75", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ce1966f-9d", "ovs_interfaceid": "2ce1966f-9d77-419b-8508-acc03ab5d271", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.925132] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7c7325-c608-418c-bf33-3c4f8da36569 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "83616524-2530-4398-9019-148e0658720d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.677s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.062566] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528e478e-2ea7-c1b9-ef85-2d88c68a4d9a, 'name': SearchDatastore_Task, 'duration_secs': 0.022946} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.063573] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e648da4-298a-4f39-a696-72584a125145 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.071538] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 836.071538] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bf2400-ad0a-7ad4-dbb1-992806d240c3" [ 836.071538] env[61911]: _type = "Task" [ 836.071538] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.080359] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bf2400-ad0a-7ad4-dbb1-992806d240c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.092949] env[61911]: DEBUG nova.network.neutron [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.164906] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.165711] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.122s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.167375] env[61911]: INFO nova.compute.claims [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.178761] env[61911]: DEBUG nova.network.neutron [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.348504] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.348504] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7deaf6b0-643a-4c82-96cb-b39fd51d6295 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.356067] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 836.356067] env[61911]: value = "task-1251071" [ 836.356067] env[61911]: _type = "Task" [ 836.356067] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.365919] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.399917] env[61911]: DEBUG nova.compute.manager [None req-2827d893-5903-4c2c-9c68-f3f7ff41a261 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Instance disappeared during snapshot {{(pid=61911) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 836.421689] env[61911]: DEBUG oslo_concurrency.lockutils [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] Releasing lock "refresh_cache-613544bc-92b9-4c8f-8a7b-f02205f347ff" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.422020] env[61911]: DEBUG nova.compute.manager [req-3789a14e-0045-4b19-bbc2-96bdc0d28ce5 req-25ebde1e-ed39-4cab-bc8c-d7258608d2b9 service nova] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Received event network-vif-deleted-8835f0fb-e13b-4847-b5d2-15a1e49d7785 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 836.549273] env[61911]: DEBUG nova.compute.manager [None req-2827d893-5903-4c2c-9c68-f3f7ff41a261 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Found 0 images (rotation: 2) {{(pid=61911) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4923}} [ 836.564622] env[61911]: DEBUG oslo_concurrency.lockutils [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "28159e75-9fe9-44c7-b5c9-534902cecbef" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.564956] env[61911]: DEBUG oslo_concurrency.lockutils [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.565922] env[61911]: INFO nova.compute.manager [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Attaching volume 4c9c79e1-b161-4bda-8c98-321d0adcdd98 to /dev/sdb [ 836.583856] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bf2400-ad0a-7ad4-dbb1-992806d240c3, 'name': SearchDatastore_Task, 'duration_secs': 0.019497} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.584013] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.584298] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 613544bc-92b9-4c8f-8a7b-f02205f347ff/613544bc-92b9-4c8f-8a7b-f02205f347ff.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.584573] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a835ee1e-c4b7-4328-b17f-1f7e4af3e0e0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.592161] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 836.592161] env[61911]: value = "task-1251072" [ 836.592161] env[61911]: _type = "Task" [ 836.592161] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.604705] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251072, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.608894] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8068f4-7e9f-4473-82f5-20f87f9495b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.616274] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8b5907-ae1f-460b-b521-b251e4097862 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.631761] env[61911]: DEBUG nova.virt.block_device [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Updating existing volume attachment record: 5ef6e2ea-34df-4e58-ba6b-e585a3b6c62a {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 836.670382] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "5ef69361-f4b0-404a-8cf4-b130f9408483" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.670666] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "5ef69361-f4b0-404a-8cf4-b130f9408483" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.682459] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Releasing lock "refresh_cache-65d310c6-d438-4c3d-bd44-f3b51123fe93" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.683145] env[61911]: DEBUG nova.compute.manager [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 836.683497] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 836.685178] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7cb84c-8636-4254-adea-0cc32df391e1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.695335] env[61911]: INFO nova.compute.manager [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Rescuing [ 836.695627] env[61911]: DEBUG oslo_concurrency.lockutils [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.695798] env[61911]: DEBUG oslo_concurrency.lockutils [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.695964] env[61911]: DEBUG nova.network.neutron [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.697154] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.697662] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e665147e-2fdb-4e77-ab03-c63e70e067c9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.704171] env[61911]: DEBUG oslo_vmware.api [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 836.704171] env[61911]: value = "task-1251073" [ 836.704171] env[61911]: _type = "Task" [ 836.704171] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.715459] env[61911]: DEBUG oslo_vmware.api [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251073, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.867131] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251071, 'name': PowerOffVM_Task, 'duration_secs': 0.217541} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.870311] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 836.871274] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674cd67b-e4dd-48f3-ac64-cc73af9ac832 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.897238] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a4c5e7-e0dc-4d10-8347-a665a9d5cf06 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.929791] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.930120] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfe9fe50-0e06-4726-a1c5-f5645d93a7c6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.941347] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 836.941347] env[61911]: value = "task-1251076" [ 836.941347] env[61911]: _type = "Task" [ 836.941347] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.954889] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 836.955658] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.955658] env[61911]: DEBUG oslo_concurrency.lockutils [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.955859] env[61911]: DEBUG oslo_concurrency.lockutils [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.955921] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.956206] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf5f3a77-e27f-4c00-9509-9ca72077c210 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.966902] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.967239] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.968201] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e962082-3810-49de-a37a-f858786edc99 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.974558] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 836.974558] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a24897-c2a9-5c17-c64a-fbdfe6cb3c7d" [ 836.974558] env[61911]: _type = "Task" [ 836.974558] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.983980] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a24897-c2a9-5c17-c64a-fbdfe6cb3c7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.103117] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251072, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.175069] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "5ef69361-f4b0-404a-8cf4-b130f9408483" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.175731] env[61911]: DEBUG nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 837.220318] env[61911]: DEBUG oslo_vmware.api [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251073, 'name': PowerOffVM_Task, 'duration_secs': 0.192306} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.220800] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 837.220800] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 837.221118] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c20d075c-3a0e-471d-95f2-e07f4804bd21 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.250961] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 837.251434] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 837.251734] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Deleting the datastore file [datastore1] 65d310c6-d438-4c3d-bd44-f3b51123fe93 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.252140] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef22f5dd-1008-4d00-acdc-845b83555754 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.263126] env[61911]: DEBUG oslo_vmware.api [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for the task: (returnval){ [ 837.263126] env[61911]: value = "task-1251079" [ 837.263126] env[61911]: _type = "Task" [ 837.263126] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.277980] env[61911]: DEBUG oslo_vmware.api [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251079, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.485283] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a24897-c2a9-5c17-c64a-fbdfe6cb3c7d, 'name': SearchDatastore_Task, 'duration_secs': 0.024583} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.489638] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f18afdfb-49a1-48d0-afb5-1cb0d634c795 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.495013] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 837.495013] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525cdf67-4343-d54e-1747-139503e962e7" [ 837.495013] env[61911]: _type = "Task" [ 837.495013] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.506991] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525cdf67-4343-d54e-1747-139503e962e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.549674] env[61911]: DEBUG nova.network.neutron [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Updating instance_info_cache with network_info: [{"id": "28654114-34d7-4037-8226-713299d2e033", "address": "fa:16:3e:42:92:99", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28654114-34", "ovs_interfaceid": "28654114-34d7-4037-8226-713299d2e033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.584524] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef19af2-bfd1-4af0-9783-c6396e4685a6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.593059] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d65ab4-fd9f-4203-b3ae-3b0480288b93 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.605442] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251072, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.776434} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.638337] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 613544bc-92b9-4c8f-8a7b-f02205f347ff/613544bc-92b9-4c8f-8a7b-f02205f347ff.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.638572] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.639409] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43c43d5e-3c7b-4c28-b904-5d39e1636064 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.641938] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7175335-8aa9-4569-9bac-187ddba0d667 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.650709] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee562bd-207b-4e16-91a3-20da87ff00f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.655148] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 837.655148] env[61911]: value = "task-1251080" [ 837.655148] env[61911]: _type = "Task" [ 837.655148] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.666947] env[61911]: DEBUG nova.compute.provider_tree [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.672907] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.684308] env[61911]: DEBUG nova.compute.utils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.685904] env[61911]: DEBUG nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 837.686103] env[61911]: DEBUG nova.network.neutron [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.765504] env[61911]: DEBUG nova.policy [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b65c27beef9c410ea62e8d94547059ee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68947e33f79d4ca1bf36a892623b2bff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.776396] env[61911]: DEBUG oslo_vmware.api [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Task: {'id': task-1251079, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.353544} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.777033] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.777033] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 837.777201] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 837.777495] env[61911]: INFO nova.compute.manager [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Took 1.09 seconds to destroy the instance on the hypervisor. [ 837.777615] env[61911]: DEBUG oslo.service.loopingcall [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.777839] env[61911]: DEBUG nova.compute.manager [-] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 837.777901] env[61911]: DEBUG nova.network.neutron [-] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 837.807474] env[61911]: DEBUG nova.network.neutron [-] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.009028] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525cdf67-4343-d54e-1747-139503e962e7, 'name': SearchDatastore_Task, 'duration_secs': 0.010104} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.009028] env[61911]: DEBUG oslo_concurrency.lockutils [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.009028] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. {{(pid=61911) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 838.009028] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbbdef76-07b0-46f8-8a3c-337d96b68371 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.014145] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 838.014145] env[61911]: value = "task-1251081" [ 838.014145] env[61911]: _type = "Task" [ 838.014145] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.023233] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.054433] env[61911]: DEBUG oslo_concurrency.lockutils [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "refresh_cache-83616524-2530-4398-9019-148e0658720d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.120057] env[61911]: DEBUG nova.network.neutron [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Successfully created port: bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.167402] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074063} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.167747] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.168688] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29fc039-5569-4967-91a3-7510ec8d3f30 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.172182] env[61911]: DEBUG nova.scheduler.client.report [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 838.196794] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 613544bc-92b9-4c8f-8a7b-f02205f347ff/613544bc-92b9-4c8f-8a7b-f02205f347ff.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.198082] env[61911]: DEBUG nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 838.200511] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d81296f-f71a-4c6f-8ee5-7e8772e35cd4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.222773] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 838.222773] env[61911]: value = "task-1251082" [ 838.222773] env[61911]: _type = "Task" [ 838.222773] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.232724] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.309856] env[61911]: DEBUG nova.network.neutron [-] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.525669] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251081, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.679087] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.679087] env[61911]: DEBUG nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 838.683329] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.582s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.685172] env[61911]: DEBUG nova.objects.instance [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lazy-loading 'resources' on Instance uuid 28da8a83-a004-4baf-a744-bcc260032afd {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.738568] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251082, 'name': ReconfigVM_Task, 'duration_secs': 0.481516} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.741028] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 613544bc-92b9-4c8f-8a7b-f02205f347ff/613544bc-92b9-4c8f-8a7b-f02205f347ff.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.741028] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57d41793-05cd-4004-be29-472e6e7aac44 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.749323] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 838.749323] env[61911]: value = "task-1251083" [ 838.749323] env[61911]: _type = "Task" [ 838.749323] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.759117] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251083, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.812711] env[61911]: INFO nova.compute.manager [-] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Took 1.03 seconds to deallocate network for instance. [ 839.024424] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.83156} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.024718] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. [ 839.025499] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5349397d-1b48-4bc4-a394-f718fa494a71 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.049778] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.050062] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8495468-ec43-47c8-a6ce-53bbb158f8de {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.067094] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 839.067094] env[61911]: value = "task-1251084" [ 839.067094] env[61911]: _type = "Task" [ 839.067094] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.074641] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.185280] env[61911]: DEBUG nova.compute.utils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.186832] env[61911]: DEBUG nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 839.186954] env[61911]: DEBUG nova.network.neutron [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.223424] env[61911]: DEBUG nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 839.233708] env[61911]: DEBUG nova.policy [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d850b9eeb164e7883e19ebe4d35b643', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e00325b35c074e92a5c80bf2a2df08dd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.248789] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 839.248789] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.248956] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 839.249687] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.249687] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 839.249687] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 839.249687] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 839.250206] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 839.250206] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 839.250206] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 839.250206] env[61911]: DEBUG nova.virt.hardware [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 839.251227] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e7073d-0d1e-4f71-9456-94b367fb67f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.266885] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299943c9-b4f2-42a3-a782-b5a530f58676 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.271971] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251083, 'name': Rename_Task, 'duration_secs': 0.281541} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.273450] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.274210] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75bca1ac-883f-4fb1-aa89-e4e1a8c80ebd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.288312] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 839.288312] env[61911]: value = "task-1251085" [ 839.288312] env[61911]: _type = "Task" [ 839.288312] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.295609] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251085, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.319531] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.480136] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50c557f-87bf-4bac-bf08-e287718f77f4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.489037] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f511714-2706-4e99-8fa9-1d558b6d5bfd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.520119] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81797502-086f-409c-bdf3-6dbabe0917fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.527538] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85accc65-7d25-4708-9fcb-fcde8a5eb360 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.540332] env[61911]: DEBUG nova.compute.provider_tree [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.580208] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251084, 'name': ReconfigVM_Task, 'duration_secs': 0.331485} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.580644] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Reconfigured VM instance instance-0000003e to attach disk [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.581548] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b7adab-91a1-48af-9edb-0ff1ba9c93c0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.602417] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.611796] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5daca71c-335f-44a3-9779-be2dfe58a6e5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.614128] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35f4a0e3-c4ea-4c4c-8b64-c397660a709f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.631451] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 839.631451] env[61911]: value = "task-1251086" [ 839.631451] env[61911]: _type = "Task" [ 839.631451] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.633028] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 839.633028] env[61911]: value = "task-1251087" [ 839.633028] env[61911]: _type = "Task" [ 839.633028] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.644078] env[61911]: DEBUG nova.network.neutron [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Successfully created port: 21db17c6-a0d9-48bc-96ed-b51ba77e6d40 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.655747] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.656065] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251087, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.691087] env[61911]: DEBUG nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 839.798719] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251085, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.807597] env[61911]: DEBUG nova.compute.manager [req-af242caf-20f0-4b23-b53f-38214ec7a1dc req-79f3f6bb-e395-4d49-baa5-3d0aec493d62 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Received event network-vif-plugged-bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 839.807824] env[61911]: DEBUG oslo_concurrency.lockutils [req-af242caf-20f0-4b23-b53f-38214ec7a1dc req-79f3f6bb-e395-4d49-baa5-3d0aec493d62 service nova] Acquiring lock "6e100d92-eb95-4b9b-b236-34ff94051811-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.808073] env[61911]: DEBUG oslo_concurrency.lockutils [req-af242caf-20f0-4b23-b53f-38214ec7a1dc req-79f3f6bb-e395-4d49-baa5-3d0aec493d62 service nova] Lock "6e100d92-eb95-4b9b-b236-34ff94051811-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.808309] env[61911]: DEBUG oslo_concurrency.lockutils [req-af242caf-20f0-4b23-b53f-38214ec7a1dc req-79f3f6bb-e395-4d49-baa5-3d0aec493d62 service nova] Lock "6e100d92-eb95-4b9b-b236-34ff94051811-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.808426] env[61911]: DEBUG nova.compute.manager [req-af242caf-20f0-4b23-b53f-38214ec7a1dc req-79f3f6bb-e395-4d49-baa5-3d0aec493d62 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] No waiting events found dispatching network-vif-plugged-bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 839.808597] env[61911]: WARNING nova.compute.manager [req-af242caf-20f0-4b23-b53f-38214ec7a1dc req-79f3f6bb-e395-4d49-baa5-3d0aec493d62 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Received unexpected event network-vif-plugged-bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 for instance with vm_state building and task_state spawning. [ 840.043103] env[61911]: DEBUG nova.scheduler.client.report [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 840.150568] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251087, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.150901] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251086, 'name': PowerOffVM_Task, 'duration_secs': 0.27947} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.151172] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.151983] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb603d3-51b0-48c5-a709-50e09cd6e36d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.173126] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9faea5-d5c8-4158-a9ad-d70251747e19 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.217870] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.220087] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4352bd6-80f9-4ef7-bae9-c692b85c3d67 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.229478] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 840.229478] env[61911]: value = "task-1251088" [ 840.229478] env[61911]: _type = "Task" [ 840.229478] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.241329] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 840.241562] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 840.242285] env[61911]: DEBUG oslo_concurrency.lockutils [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.242478] env[61911]: DEBUG oslo_concurrency.lockutils [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.242839] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 840.243369] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4bb7f385-6be4-401a-9f22-f117a3d46e66 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.253871] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 840.254107] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 840.254928] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19a0b417-6cb9-4f51-9f78-2f49abde620f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.260513] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 840.260513] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520d18a4-a8c6-8dc3-2d9b-ebeaefc9cde2" [ 840.260513] env[61911]: _type = "Task" [ 840.260513] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.268821] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520d18a4-a8c6-8dc3-2d9b-ebeaefc9cde2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.300120] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251085, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.410909] env[61911]: DEBUG nova.network.neutron [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Successfully updated port: bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.529714] env[61911]: DEBUG nova.compute.manager [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Received event network-changed-bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 840.529972] env[61911]: DEBUG nova.compute.manager [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Refreshing instance network info cache due to event network-changed-bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 840.530122] env[61911]: DEBUG oslo_concurrency.lockutils [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] Acquiring lock "refresh_cache-6e100d92-eb95-4b9b-b236-34ff94051811" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.530273] env[61911]: DEBUG oslo_concurrency.lockutils [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] Acquired lock "refresh_cache-6e100d92-eb95-4b9b-b236-34ff94051811" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.530437] env[61911]: DEBUG nova.network.neutron [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Refreshing network info cache for port bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.548105] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.550663] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.066s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.551627] env[61911]: INFO nova.compute.claims [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.570436] env[61911]: INFO nova.scheduler.client.report [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Deleted allocations for instance 28da8a83-a004-4baf-a744-bcc260032afd [ 840.647886] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251087, 'name': ReconfigVM_Task, 'duration_secs': 0.597225} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.648289] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.648560] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33ebb8ca-af2d-4db8-afde-a011b6916f5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.655817] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 840.655817] env[61911]: value = "task-1251089" [ 840.655817] env[61911]: _type = "Task" [ 840.655817] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.663423] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.701726] env[61911]: DEBUG nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 840.725791] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 840.726133] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.726326] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 840.726520] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.726679] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 840.726834] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 840.727056] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 840.727234] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 840.727407] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 840.727575] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 840.727753] env[61911]: DEBUG nova.virt.hardware [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 840.728638] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5258d05-9947-4499-9478-5c28b10fcb5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.737137] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4191404e-a024-4e6c-8e39-0bea03f8b5ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.769994] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520d18a4-a8c6-8dc3-2d9b-ebeaefc9cde2, 'name': SearchDatastore_Task, 'duration_secs': 0.010955} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.770786] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b7b31b5-e725-4e94-9a56-ce53c23f4f9a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.775599] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 840.775599] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5245f7f6-db90-4fcc-8973-e8e1bb63325b" [ 840.775599] env[61911]: _type = "Task" [ 840.775599] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.782706] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5245f7f6-db90-4fcc-8973-e8e1bb63325b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.797505] env[61911]: DEBUG oslo_vmware.api [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251085, 'name': PowerOnVM_Task, 'duration_secs': 1.096547} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.797777] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.798014] env[61911]: INFO nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Took 8.85 seconds to spawn the instance on the hypervisor. [ 840.798221] env[61911]: DEBUG nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 840.799270] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9829968-9e25-41e1-b41c-1829191afedb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.918850] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "refresh_cache-6e100d92-eb95-4b9b-b236-34ff94051811" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.068091] env[61911]: DEBUG nova.network.neutron [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.077603] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a20eb734-0673-4cc4-8b36-932efe9dd9c6 tempest-ServerMetadataNegativeTestJSON-1194883294 tempest-ServerMetadataNegativeTestJSON-1194883294-project-member] Lock "28da8a83-a004-4baf-a744-bcc260032afd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.313s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.146134] env[61911]: DEBUG nova.network.neutron [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.166782] env[61911]: DEBUG oslo_vmware.api [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251089, 'name': PowerOnVM_Task, 'duration_secs': 0.489713} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.167070] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.169850] env[61911]: DEBUG nova.compute.manager [None req-257ba019-8289-47de-92b2-f07934748235 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 841.170647] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7421b6c-78f9-4842-8b83-978cec010ad1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.286469] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5245f7f6-db90-4fcc-8973-e8e1bb63325b, 'name': SearchDatastore_Task, 'duration_secs': 0.008767} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.286813] env[61911]: DEBUG oslo_concurrency.lockutils [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.287150] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 83616524-2530-4398-9019-148e0658720d/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. {{(pid=61911) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 841.287493] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30b8e9ae-a50b-4277-9d3d-011a0bcdf348 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.294283] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 841.294283] env[61911]: value = "task-1251091" [ 841.294283] env[61911]: _type = "Task" [ 841.294283] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.302854] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251091, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.318892] env[61911]: INFO nova.compute.manager [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Took 27.73 seconds to build instance. [ 841.649312] env[61911]: DEBUG oslo_concurrency.lockutils [req-4df137a1-6475-4a6a-884d-9f7250e715ec req-2c049861-c696-40d5-88d1-7ab349a9aa53 service nova] Releasing lock "refresh_cache-6e100d92-eb95-4b9b-b236-34ff94051811" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.649705] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquired lock "refresh_cache-6e100d92-eb95-4b9b-b236-34ff94051811" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.649862] env[61911]: DEBUG nova.network.neutron [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.806381] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251091, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42741} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.806726] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 83616524-2530-4398-9019-148e0658720d/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. [ 841.807873] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c002dc2-f49c-4140-aa32-f92b0ffea9eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.832210] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7a52ddaf-758e-4b2f-933e-5f11df3bbfd6 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.694s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.841035] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 83616524-2530-4398-9019-148e0658720d/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.844460] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afdb42db-431f-46c6-8ca8-0efd752204f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.865300] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 841.865300] env[61911]: value = "task-1251092" [ 841.865300] env[61911]: _type = "Task" [ 841.865300] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.878349] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251092, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.898664] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cead9df5-3bc6-4584-a246-861c3ab0a938 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.905576] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f550ede-5ab1-4b55-b961-93a1add54328 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.936421] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e7d9b4-4910-4529-9fa0-5001bf7181d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.943993] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3cc575-efb9-4509-b452-40e45036a731 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.957492] env[61911]: DEBUG nova.compute.provider_tree [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.002432] env[61911]: DEBUG nova.network.neutron [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Successfully updated port: 21db17c6-a0d9-48bc-96ed-b51ba77e6d40 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.194520] env[61911]: DEBUG nova.network.neutron [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.379254] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251092, 'name': ReconfigVM_Task, 'duration_secs': 0.372347} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.379254] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 83616524-2530-4398-9019-148e0658720d/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.379620] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4eb940-060e-44b4-a1b7-3a5000422c29 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.406674] env[61911]: DEBUG nova.network.neutron [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Updating instance_info_cache with network_info: [{"id": "bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8", "address": "fa:16:3e:1e:18:34", "network": {"id": "25edbf44-a733-4972-bc26-23802a90a6bb", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-966580787-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68947e33f79d4ca1bf36a892623b2bff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdaa8bb9-f6", "ovs_interfaceid": "bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.408015] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af86f571-3897-4423-a0cd-77d07f714550 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.424178] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 842.424178] env[61911]: value = "task-1251093" [ 842.424178] env[61911]: _type = "Task" [ 842.424178] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.432781] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251093, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.461355] env[61911]: DEBUG nova.scheduler.client.report [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 842.505214] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "refresh_cache-4b113b46-1da2-49ad-a0bb-c01b56c70f00" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.505387] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "refresh_cache-4b113b46-1da2-49ad-a0bb-c01b56c70f00" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.506111] env[61911]: DEBUG nova.network.neutron [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.518176] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.518423] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.554590] env[61911]: DEBUG nova.compute.manager [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Received event network-vif-plugged-21db17c6-a0d9-48bc-96ed-b51ba77e6d40 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 842.554816] env[61911]: DEBUG oslo_concurrency.lockutils [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] Acquiring lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.555045] env[61911]: DEBUG oslo_concurrency.lockutils [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.555191] env[61911]: DEBUG oslo_concurrency.lockutils [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.555358] env[61911]: DEBUG nova.compute.manager [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] No waiting events found dispatching network-vif-plugged-21db17c6-a0d9-48bc-96ed-b51ba77e6d40 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 842.555520] env[61911]: WARNING nova.compute.manager [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Received unexpected event network-vif-plugged-21db17c6-a0d9-48bc-96ed-b51ba77e6d40 for instance with vm_state building and task_state spawning. [ 842.555680] env[61911]: DEBUG nova.compute.manager [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Received event network-changed-21db17c6-a0d9-48bc-96ed-b51ba77e6d40 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 842.555893] env[61911]: DEBUG nova.compute.manager [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Refreshing instance network info cache due to event network-changed-21db17c6-a0d9-48bc-96ed-b51ba77e6d40. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 842.556129] env[61911]: DEBUG oslo_concurrency.lockutils [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] Acquiring lock "refresh_cache-4b113b46-1da2-49ad-a0bb-c01b56c70f00" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.919310] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Releasing lock "refresh_cache-6e100d92-eb95-4b9b-b236-34ff94051811" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.919658] env[61911]: DEBUG nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Instance network_info: |[{"id": "bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8", "address": "fa:16:3e:1e:18:34", "network": {"id": "25edbf44-a733-4972-bc26-23802a90a6bb", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-966580787-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "68947e33f79d4ca1bf36a892623b2bff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f", "external-id": "nsx-vlan-transportzone-584", "segmentation_id": 584, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdaa8bb9-f6", "ovs_interfaceid": "bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 842.920097] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:18:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '196c2dd2-7ffc-4f7d-9c93-e1ef0a6a3a9f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.927520] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Creating folder: Project (68947e33f79d4ca1bf36a892623b2bff). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.927869] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21b2295c-a607-45c8-a73a-7ebab5abd890 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.938186] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251093, 'name': ReconfigVM_Task, 'duration_secs': 0.327067} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.938468] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.938723] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e854c28-7e26-4992-a9cc-b924c5dc2ae2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.941312] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Created folder: Project (68947e33f79d4ca1bf36a892623b2bff) in parent group-v269521. [ 842.941497] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Creating folder: Instances. Parent ref: group-v269595. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.942045] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-392accde-5795-4d25-8d49-8d476d0f1032 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.946327] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 842.946327] env[61911]: value = "task-1251095" [ 842.946327] env[61911]: _type = "Task" [ 842.946327] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.952159] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Created folder: Instances in parent group-v269595. [ 842.952159] env[61911]: DEBUG oslo.service.loopingcall [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.952159] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.952159] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02153d81-28d2-4933-81f5-8b633564525e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.970978] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.972018] env[61911]: DEBUG nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 842.974343] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251095, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.975065] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.806s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.976465] env[61911]: INFO nova.compute.claims [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.984519] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.984519] env[61911]: value = "task-1251097" [ 842.984519] env[61911]: _type = "Task" [ 842.984519] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.992922] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251097, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.020429] env[61911]: DEBUG nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 843.129220] env[61911]: DEBUG nova.network.neutron [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.183338] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 843.183673] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269594', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'name': 'volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '28159e75-9fe9-44c7-b5c9-534902cecbef', 'attached_at': '', 'detached_at': '', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'serial': '4c9c79e1-b161-4bda-8c98-321d0adcdd98'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 843.184771] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89372039-f991-420b-8e1b-b37a25e9867f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.206763] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffec347-1e6c-4930-b483-69939641d8a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.239010] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98/volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.242624] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f449c8a1-6b3b-4e28-8d5e-ba1efb5a9378 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.266058] env[61911]: DEBUG oslo_vmware.api [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 843.266058] env[61911]: value = "task-1251098" [ 843.266058] env[61911]: _type = "Task" [ 843.266058] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.274539] env[61911]: DEBUG oslo_vmware.api [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251098, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.446335] env[61911]: DEBUG nova.network.neutron [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Updating instance_info_cache with network_info: [{"id": "21db17c6-a0d9-48bc-96ed-b51ba77e6d40", "address": "fa:16:3e:0a:37:c9", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21db17c6-a0", "ovs_interfaceid": "21db17c6-a0d9-48bc-96ed-b51ba77e6d40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.462183] env[61911]: DEBUG oslo_vmware.api [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251095, 'name': PowerOnVM_Task, 'duration_secs': 0.418189} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.462548] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.468399] env[61911]: DEBUG nova.compute.manager [None req-72b51e41-68f7-4a7c-97a5-4a10b9d60cfd tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 843.469508] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2268dad-997f-4986-80cc-d18d511e021e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.477510] env[61911]: DEBUG nova.compute.utils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.480939] env[61911]: DEBUG nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 843.481134] env[61911]: DEBUG nova.network.neutron [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.500256] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251097, 'name': CreateVM_Task, 'duration_secs': 0.442647} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.500449] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.501173] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.501402] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.501741] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.502139] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6c8d92d-5e8c-4932-8cd9-893a6d6f8573 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.508140] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 843.508140] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c3b582-ee53-15ff-b42c-771f5c9f8f22" [ 843.508140] env[61911]: _type = "Task" [ 843.508140] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.517857] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c3b582-ee53-15ff-b42c-771f5c9f8f22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.536289] env[61911]: DEBUG nova.policy [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b034026eede4f5ca4757b552f7dca51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5be8ec1d59a14d618906ec32d0e134e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.544649] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.778087] env[61911]: DEBUG oslo_vmware.api [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.908022] env[61911]: DEBUG nova.network.neutron [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Successfully created port: 47a25084-ef5a-4059-92b8-d93d72091857 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.956403] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "refresh_cache-4b113b46-1da2-49ad-a0bb-c01b56c70f00" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.956403] env[61911]: DEBUG nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Instance network_info: |[{"id": "21db17c6-a0d9-48bc-96ed-b51ba77e6d40", "address": "fa:16:3e:0a:37:c9", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21db17c6-a0", "ovs_interfaceid": "21db17c6-a0d9-48bc-96ed-b51ba77e6d40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 843.956403] env[61911]: DEBUG oslo_concurrency.lockutils [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] Acquired lock "refresh_cache-4b113b46-1da2-49ad-a0bb-c01b56c70f00" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.956403] env[61911]: DEBUG nova.network.neutron [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Refreshing network info cache for port 21db17c6-a0d9-48bc-96ed-b51ba77e6d40 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 843.957124] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:37:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21db17c6-a0d9-48bc-96ed-b51ba77e6d40', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.971040] env[61911]: DEBUG oslo.service.loopingcall [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.971129] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.971338] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af007e9b-fd39-4f45-826d-4c3c1d1039a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.991551] env[61911]: DEBUG nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 844.001802] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.001802] env[61911]: value = "task-1251099" [ 844.001802] env[61911]: _type = "Task" [ 844.001802] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.011719] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251099, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.025126] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c3b582-ee53-15ff-b42c-771f5c9f8f22, 'name': SearchDatastore_Task, 'duration_secs': 0.010851} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.025126] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.025126] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.025126] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.025275] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.025458] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.025727] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-212e6699-122e-4b0f-813d-acc07ccd6052 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.048529] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.048720] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.049508] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da130702-55bc-4630-9145-b41e0181bc19 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.055150] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 844.055150] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52603f51-a60f-0250-acfc-ff1c50eedefb" [ 844.055150] env[61911]: _type = "Task" [ 844.055150] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.063408] env[61911]: INFO nova.compute.manager [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Rescuing [ 844.063408] env[61911]: DEBUG oslo_concurrency.lockutils [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.063408] env[61911]: DEBUG oslo_concurrency.lockutils [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.063590] env[61911]: DEBUG nova.network.neutron [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.064731] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52603f51-a60f-0250-acfc-ff1c50eedefb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.278031] env[61911]: DEBUG oslo_vmware.api [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251098, 'name': ReconfigVM_Task, 'duration_secs': 0.624241} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.286942] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Reconfigured VM instance instance-00000033 to attach disk [datastore1] volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98/volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.292262] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1b09e2b-66af-4650-a710-7241964e40d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.312149] env[61911]: DEBUG oslo_vmware.api [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 844.312149] env[61911]: value = "task-1251100" [ 844.312149] env[61911]: _type = "Task" [ 844.312149] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.322736] env[61911]: DEBUG oslo_vmware.api [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251100, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.415225] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba31758-c852-46fe-b00b-319749eeea94 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.423815] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385570c3-f993-492a-bd7e-51f3da5bd089 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.461029] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea282d9-5a7b-41e9-8f26-efb4537a039c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.469785] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5587250-e76d-4c1c-bac9-e196cbc73827 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.491842] env[61911]: DEBUG nova.compute.provider_tree [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.517089] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251099, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.565368] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52603f51-a60f-0250-acfc-ff1c50eedefb, 'name': SearchDatastore_Task, 'duration_secs': 0.017941} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.567978] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34ae77b0-b9f3-4328-8537-7cfa8955d2af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.573685] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 844.573685] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525b6d27-110f-0c56-4475-310f0050af5e" [ 844.573685] env[61911]: _type = "Task" [ 844.573685] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.585125] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525b6d27-110f-0c56-4475-310f0050af5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.781392] env[61911]: DEBUG nova.network.neutron [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Updating instance_info_cache with network_info: [{"id": "15e642fd-30d2-41c2-98f7-0d19e729041e", "address": "fa:16:3e:8e:99:bd", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e642fd-30", "ovs_interfaceid": "15e642fd-30d2-41c2-98f7-0d19e729041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.824405] env[61911]: DEBUG oslo_vmware.api [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251100, 'name': ReconfigVM_Task, 'duration_secs': 0.160603} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.824751] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269594', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'name': 'volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '28159e75-9fe9-44c7-b5c9-534902cecbef', 'attached_at': '', 'detached_at': '', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'serial': '4c9c79e1-b161-4bda-8c98-321d0adcdd98'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 844.911463] env[61911]: DEBUG nova.network.neutron [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Updated VIF entry in instance network info cache for port 21db17c6-a0d9-48bc-96ed-b51ba77e6d40. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.912635] env[61911]: DEBUG nova.network.neutron [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Updating instance_info_cache with network_info: [{"id": "21db17c6-a0d9-48bc-96ed-b51ba77e6d40", "address": "fa:16:3e:0a:37:c9", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21db17c6-a0", "ovs_interfaceid": "21db17c6-a0d9-48bc-96ed-b51ba77e6d40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.996059] env[61911]: DEBUG nova.scheduler.client.report [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 845.007552] env[61911]: DEBUG nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 845.013385] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251099, 'name': CreateVM_Task, 'duration_secs': 0.527851} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.014836] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.015753] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.015753] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.016123] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.016619] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5436aa1-e039-4b87-84a5-a771c635b5cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.022164] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 845.022164] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527274ce-aca3-a80a-c953-1c602a183139" [ 845.022164] env[61911]: _type = "Task" [ 845.022164] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.032277] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527274ce-aca3-a80a-c953-1c602a183139, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.050558] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 845.050802] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.050966] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 845.051174] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.051325] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 845.051475] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 845.051718] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 845.051897] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 845.052200] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 845.052394] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 845.052575] env[61911]: DEBUG nova.virt.hardware [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 845.055463] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe10ff3-165a-48aa-b4fe-107eb5345515 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.062664] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "1058f8d2-0979-4423-802d-52d91d5a2ff7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.062931] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.067780] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ac3511-d4d2-4b80-8218-cdcf75c419da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.091172] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525b6d27-110f-0c56-4475-310f0050af5e, 'name': SearchDatastore_Task, 'duration_secs': 0.01185} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.091446] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.091715] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 6e100d92-eb95-4b9b-b236-34ff94051811/6e100d92-eb95-4b9b-b236-34ff94051811.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.091968] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-035fec95-dd42-429c-9117-5e66c19fe5d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.098876] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 845.098876] env[61911]: value = "task-1251101" [ 845.098876] env[61911]: _type = "Task" [ 845.098876] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.108057] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251101, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.283557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.414521] env[61911]: DEBUG oslo_concurrency.lockutils [req-abd8f29d-113d-43d2-8422-7a7aae1aaca3 req-ecde2216-68a8-4891-bd43-fc58b78459e5 service nova] Releasing lock "refresh_cache-4b113b46-1da2-49ad-a0bb-c01b56c70f00" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.501362] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.501982] env[61911]: DEBUG nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 845.505032] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.862s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.505231] env[61911]: DEBUG nova.objects.instance [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lazy-loading 'resources' on Instance uuid 35cf6a24-3cc0-40bd-92da-798a56b4e2c9 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.540507] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527274ce-aca3-a80a-c953-1c602a183139, 'name': SearchDatastore_Task, 'duration_secs': 0.011995} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.541047] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.541144] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.541475] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.541626] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.541809] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.542190] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f623c951-ecf7-4be6-83c4-66a94559a373 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.556416] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.556416] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.557609] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef0de361-18a3-463a-9431-b5701fb08a8b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.564917] env[61911]: DEBUG nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 845.569690] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 845.569690] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524d9892-30f6-1c7a-88e9-51662d995ad3" [ 845.569690] env[61911]: _type = "Task" [ 845.569690] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.582388] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524d9892-30f6-1c7a-88e9-51662d995ad3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.610032] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251101, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.616548] env[61911]: DEBUG nova.compute.manager [req-b268ba14-c60f-4bae-8d9b-2b0a362d1d12 req-4d94118a-32b2-4bc5-ac18-f256f4ddb906 service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Received event network-vif-plugged-47a25084-ef5a-4059-92b8-d93d72091857 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 845.616827] env[61911]: DEBUG oslo_concurrency.lockutils [req-b268ba14-c60f-4bae-8d9b-2b0a362d1d12 req-4d94118a-32b2-4bc5-ac18-f256f4ddb906 service nova] Acquiring lock "9230cd85-960d-427c-b69c-d19ac2bcfe35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.617096] env[61911]: DEBUG oslo_concurrency.lockutils [req-b268ba14-c60f-4bae-8d9b-2b0a362d1d12 req-4d94118a-32b2-4bc5-ac18-f256f4ddb906 service nova] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.617320] env[61911]: DEBUG oslo_concurrency.lockutils [req-b268ba14-c60f-4bae-8d9b-2b0a362d1d12 req-4d94118a-32b2-4bc5-ac18-f256f4ddb906 service nova] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.617457] env[61911]: DEBUG nova.compute.manager [req-b268ba14-c60f-4bae-8d9b-2b0a362d1d12 req-4d94118a-32b2-4bc5-ac18-f256f4ddb906 service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] No waiting events found dispatching network-vif-plugged-47a25084-ef5a-4059-92b8-d93d72091857 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 845.617627] env[61911]: WARNING nova.compute.manager [req-b268ba14-c60f-4bae-8d9b-2b0a362d1d12 req-4d94118a-32b2-4bc5-ac18-f256f4ddb906 service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Received unexpected event network-vif-plugged-47a25084-ef5a-4059-92b8-d93d72091857 for instance with vm_state building and task_state spawning. [ 845.744763] env[61911]: DEBUG nova.network.neutron [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Successfully updated port: 47a25084-ef5a-4059-92b8-d93d72091857 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.867135] env[61911]: DEBUG nova.objects.instance [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lazy-loading 'flavor' on Instance uuid 28159e75-9fe9-44c7-b5c9-534902cecbef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.012120] env[61911]: DEBUG nova.compute.utils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.016059] env[61911]: DEBUG nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 846.016059] env[61911]: DEBUG nova.network.neutron [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.089513] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524d9892-30f6-1c7a-88e9-51662d995ad3, 'name': SearchDatastore_Task, 'duration_secs': 0.059464} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.089513] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4352f855-689c-4385-8c7d-4aa786c9356d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.097405] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 846.097405] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb054e-1d9e-827c-83ea-fd292a434844" [ 846.097405] env[61911]: _type = "Task" [ 846.097405] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.098880] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.112751] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb054e-1d9e-827c-83ea-fd292a434844, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.114484] env[61911]: DEBUG nova.policy [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe50c877fcef489cb798fafe73caaf8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813b3181120245e594809d096a621675', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.118934] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251101, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660452} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.121677] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 6e100d92-eb95-4b9b-b236-34ff94051811/6e100d92-eb95-4b9b-b236-34ff94051811.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.121909] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.122344] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06e33c54-4b2a-444c-90c1-7ada38af8cd9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.128679] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 846.128679] env[61911]: value = "task-1251102" [ 846.128679] env[61911]: _type = "Task" [ 846.128679] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.140677] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251102, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.247756] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-9230cd85-960d-427c-b69c-d19ac2bcfe35" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.247981] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-9230cd85-960d-427c-b69c-d19ac2bcfe35" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.248755] env[61911]: DEBUG nova.network.neutron [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.369512] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7c9a11-a39d-479b-994a-7a6f774f29f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.374851] env[61911]: DEBUG oslo_concurrency.lockutils [None req-143ef129-bbf9-4bb2-9831-5156210fa32f tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.810s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.378899] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbe7ba9-44a4-4df7-b49f-f65ce20a01dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.411677] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c80fd2-5df3-41fc-b3f7-1e137f9bc19a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.419880] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a02625-d107-47db-927b-2cf74c8698c3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.435159] env[61911]: DEBUG nova.compute.provider_tree [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.435930] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "28159e75-9fe9-44c7-b5c9-534902cecbef" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.436184] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.466205] env[61911]: DEBUG nova.network.neutron [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Successfully created port: 28133c58-2aa2-41ae-9c8b-584659387203 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.518492] env[61911]: DEBUG nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 846.609711] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb054e-1d9e-827c-83ea-fd292a434844, 'name': SearchDatastore_Task, 'duration_secs': 0.045869} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.610110] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.610279] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 4b113b46-1da2-49ad-a0bb-c01b56c70f00/4b113b46-1da2-49ad-a0bb-c01b56c70f00.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.610541] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1c13c8f-6b04-4187-8d58-112870ad3372 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.616629] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 846.616629] env[61911]: value = "task-1251103" [ 846.616629] env[61911]: _type = "Task" [ 846.616629] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.626394] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.636116] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.174162} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.636387] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.637351] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88361cd-a4fc-4efd-a892-827bc41dc45f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.658838] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 6e100d92-eb95-4b9b-b236-34ff94051811/6e100d92-eb95-4b9b-b236-34ff94051811.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.659128] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd947a2c-04b6-4fb2-9a4d-d3c31219e767 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.678040] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 846.678040] env[61911]: value = "task-1251104" [ 846.678040] env[61911]: _type = "Task" [ 846.678040] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.690027] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251104, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.819286] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.819613] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8936855-59e2-4d8c-8e45-489bf9f6718c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.826690] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 846.826690] env[61911]: value = "task-1251105" [ 846.826690] env[61911]: _type = "Task" [ 846.826690] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.834042] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.862437] env[61911]: DEBUG nova.network.neutron [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.937267] env[61911]: DEBUG nova.scheduler.client.report [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 846.949041] env[61911]: INFO nova.compute.manager [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Detaching volume 4c9c79e1-b161-4bda-8c98-321d0adcdd98 [ 846.996949] env[61911]: INFO nova.virt.block_device [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Attempting to driver detach volume 4c9c79e1-b161-4bda-8c98-321d0adcdd98 from mountpoint /dev/sdb [ 846.997237] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 846.997446] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269594', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'name': 'volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '28159e75-9fe9-44c7-b5c9-534902cecbef', 'attached_at': '', 'detached_at': '', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'serial': '4c9c79e1-b161-4bda-8c98-321d0adcdd98'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 846.998439] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796d83b4-1ec3-490a-bb4d-1e5244d3be13 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.021202] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c579b6d-d46c-48f3-b990-8b58c39506a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.032787] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7579537-5c64-43c1-9b9b-675d4f5782fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.059376] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ff6318-4962-4b41-afc8-b1da12050518 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.079424] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] The volume has not been displaced from its original location: [datastore1] volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98/volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 847.084440] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Reconfiguring VM instance instance-00000033 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 847.084685] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f3638b4-df65-4cec-afa5-4a9490467c6c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.107078] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 847.107078] env[61911]: value = "task-1251106" [ 847.107078] env[61911]: _type = "Task" [ 847.107078] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.115064] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251106, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.126299] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251103, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.185313] env[61911]: DEBUG nova.network.neutron [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Updating instance_info_cache with network_info: [{"id": "47a25084-ef5a-4059-92b8-d93d72091857", "address": "fa:16:3e:b7:bd:d8", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47a25084-ef", "ovs_interfaceid": "47a25084-ef5a-4059-92b8-d93d72091857", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.191035] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251104, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.338619] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.451997] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.947s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.454800] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.587s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.455218] env[61911]: DEBUG nova.objects.instance [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lazy-loading 'resources' on Instance uuid 43d18895-202c-4048-9435-b3484ffd4c07 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.480851] env[61911]: INFO nova.scheduler.client.report [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted allocations for instance 35cf6a24-3cc0-40bd-92da-798a56b4e2c9 [ 847.529972] env[61911]: DEBUG nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 847.557399] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 847.557794] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.558047] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 847.558330] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.558561] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 847.558731] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 847.558949] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 847.559131] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 847.559309] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 847.559480] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 847.559659] env[61911]: DEBUG nova.virt.hardware [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 847.560549] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf56ad7-722e-46fb-a5d2-7719f0cd2819 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.569306] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e2e7ea-f0b1-4ab4-8470-37e35dcf7209 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.616502] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251106, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.626407] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634123} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.626407] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 4b113b46-1da2-49ad-a0bb-c01b56c70f00/4b113b46-1da2-49ad-a0bb-c01b56c70f00.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.626407] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.626894] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9d3fb6c-e5f3-412a-83e5-d88b2198caee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.633838] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 847.633838] env[61911]: value = "task-1251107" [ 847.633838] env[61911]: _type = "Task" [ 847.633838] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.641860] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.646880] env[61911]: DEBUG nova.compute.manager [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Received event network-changed-47a25084-ef5a-4059-92b8-d93d72091857 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 847.647172] env[61911]: DEBUG nova.compute.manager [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Refreshing instance network info cache due to event network-changed-47a25084-ef5a-4059-92b8-d93d72091857. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 847.647991] env[61911]: DEBUG oslo_concurrency.lockutils [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] Acquiring lock "refresh_cache-9230cd85-960d-427c-b69c-d19ac2bcfe35" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.691259] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251104, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.691934] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-9230cd85-960d-427c-b69c-d19ac2bcfe35" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.692400] env[61911]: DEBUG nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Instance network_info: |[{"id": "47a25084-ef5a-4059-92b8-d93d72091857", "address": "fa:16:3e:b7:bd:d8", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47a25084-ef", "ovs_interfaceid": "47a25084-ef5a-4059-92b8-d93d72091857", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 847.692832] env[61911]: DEBUG oslo_concurrency.lockutils [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] Acquired lock "refresh_cache-9230cd85-960d-427c-b69c-d19ac2bcfe35" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.693153] env[61911]: DEBUG nova.network.neutron [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Refreshing network info cache for port 47a25084-ef5a-4059-92b8-d93d72091857 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.694914] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:bd:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47a25084-ef5a-4059-92b8-d93d72091857', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.707659] env[61911]: DEBUG oslo.service.loopingcall [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.709009] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 847.709351] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8bd7b152-d88d-40ed-84e7-48620d619357 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.742020] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.742020] env[61911]: value = "task-1251108" [ 847.742020] env[61911]: _type = "Task" [ 847.742020] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.753857] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.840318] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251105, 'name': PowerOffVM_Task, 'duration_secs': 0.849941} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.840538] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.842388] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ee6c7d-21cc-4bd2-acfd-4dd3878fd95a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.871221] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e44b13-8a9c-4a8c-8984-45ac71bc42f5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.906787] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.908440] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cabb288-75de-4229-ae8d-e743ec9c4320 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.915981] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 847.915981] env[61911]: value = "task-1251109" [ 847.915981] env[61911]: _type = "Task" [ 847.915981] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.929721] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 847.930121] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.930421] env[61911]: DEBUG oslo_concurrency.lockutils [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.930640] env[61911]: DEBUG oslo_concurrency.lockutils [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.930909] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.931707] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f48f2720-a091-48e3-b716-4bc371801a2f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.943999] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.944316] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.945437] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53cf40be-69e8-4876-a861-a177e68971ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.953130] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 847.953130] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5244bd5a-4dc5-3295-f46b-b3d6bd7623ec" [ 847.953130] env[61911]: _type = "Task" [ 847.953130] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.970226] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5244bd5a-4dc5-3295-f46b-b3d6bd7623ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.990805] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c795794-dafe-490c-90eb-092e8fa5c921 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "35cf6a24-3cc0-40bd-92da-798a56b4e2c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.626s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.115635] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251106, 'name': ReconfigVM_Task, 'duration_secs': 0.606682} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.118865] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Reconfigured VM instance instance-00000033 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 848.125020] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cab84336-a58c-4b0b-9d8b-6db808b45348 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.143785] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 848.143785] env[61911]: value = "task-1251110" [ 848.143785] env[61911]: _type = "Task" [ 848.143785] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.147980] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070175} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.151534] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.151968] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48df3f5e-2ebd-4efe-9445-eb5761a372f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.184024] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 4b113b46-1da2-49ad-a0bb-c01b56c70f00/4b113b46-1da2-49ad-a0bb-c01b56c70f00.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.184174] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.187233] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a108c37-2b4f-4fcb-a903-bd9336de335c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.215692] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251104, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.216051] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 848.216051] env[61911]: value = "task-1251111" [ 848.216051] env[61911]: _type = "Task" [ 848.216051] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.228188] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.254889] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.316429] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f444cf4f-857b-42af-be9a-a4a78a1c099c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.324785] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f041dc-a094-4f2c-99e8-099647689b33 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.358414] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b82aba-f361-4c85-8f3d-bbe4e1528fc6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.367468] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcb141e-d57f-48ce-af05-28fcee2b44f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.382501] env[61911]: DEBUG nova.compute.provider_tree [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.468410] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5244bd5a-4dc5-3295-f46b-b3d6bd7623ec, 'name': SearchDatastore_Task, 'duration_secs': 0.02734} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.473036] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecba929c-a771-444f-8f6e-4000f73c48e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.481484] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 848.481484] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e7d6d1-b9ff-44bd-6742-4ff17ea32e2b" [ 848.481484] env[61911]: _type = "Task" [ 848.481484] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.495569] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e7d6d1-b9ff-44bd-6742-4ff17ea32e2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.658755] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.689987] env[61911]: DEBUG nova.compute.manager [req-c15d0ee6-081c-4968-a3eb-0c8a75aad860 req-c0e81a87-2821-4c18-9ae2-d354032b8b9c service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Received event network-vif-plugged-28133c58-2aa2-41ae-9c8b-584659387203 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 848.690250] env[61911]: DEBUG oslo_concurrency.lockutils [req-c15d0ee6-081c-4968-a3eb-0c8a75aad860 req-c0e81a87-2821-4c18-9ae2-d354032b8b9c service nova] Acquiring lock "63d54f69-3d8f-45e9-b317-9d484af954b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.690474] env[61911]: DEBUG oslo_concurrency.lockutils [req-c15d0ee6-081c-4968-a3eb-0c8a75aad860 req-c0e81a87-2821-4c18-9ae2-d354032b8b9c service nova] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.690645] env[61911]: DEBUG oslo_concurrency.lockutils [req-c15d0ee6-081c-4968-a3eb-0c8a75aad860 req-c0e81a87-2821-4c18-9ae2-d354032b8b9c service nova] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.690815] env[61911]: DEBUG nova.compute.manager [req-c15d0ee6-081c-4968-a3eb-0c8a75aad860 req-c0e81a87-2821-4c18-9ae2-d354032b8b9c service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] No waiting events found dispatching network-vif-plugged-28133c58-2aa2-41ae-9c8b-584659387203 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 848.690985] env[61911]: WARNING nova.compute.manager [req-c15d0ee6-081c-4968-a3eb-0c8a75aad860 req-c0e81a87-2821-4c18-9ae2-d354032b8b9c service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Received unexpected event network-vif-plugged-28133c58-2aa2-41ae-9c8b-584659387203 for instance with vm_state building and task_state spawning. [ 848.701861] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251104, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.729510] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.751926] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.776014] env[61911]: DEBUG nova.network.neutron [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Updated VIF entry in instance network info cache for port 47a25084-ef5a-4059-92b8-d93d72091857. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 848.776435] env[61911]: DEBUG nova.network.neutron [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Updating instance_info_cache with network_info: [{"id": "47a25084-ef5a-4059-92b8-d93d72091857", "address": "fa:16:3e:b7:bd:d8", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47a25084-ef", "ovs_interfaceid": "47a25084-ef5a-4059-92b8-d93d72091857", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.885735] env[61911]: DEBUG nova.scheduler.client.report [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 848.993945] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e7d6d1-b9ff-44bd-6742-4ff17ea32e2b, 'name': SearchDatastore_Task, 'duration_secs': 0.019749} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.994295] env[61911]: DEBUG oslo_concurrency.lockutils [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.994606] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. {{(pid=61911) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 848.994915] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12120efc-c7bf-449d-b467-1efe2f53f972 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.002325] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 849.002325] env[61911]: value = "task-1251112" [ 849.002325] env[61911]: _type = "Task" [ 849.002325] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.012307] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.089245] env[61911]: DEBUG nova.network.neutron [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Successfully updated port: 28133c58-2aa2-41ae-9c8b-584659387203 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.158218] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251110, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.199133] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251104, 'name': ReconfigVM_Task, 'duration_secs': 2.46608} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.199305] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 6e100d92-eb95-4b9b-b236-34ff94051811/6e100d92-eb95-4b9b-b236-34ff94051811.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.200072] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3161689d-d892-4b9e-916d-b11a31dc3871 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.205868] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 849.205868] env[61911]: value = "task-1251113" [ 849.205868] env[61911]: _type = "Task" [ 849.205868] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.214539] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251113, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.228957] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251111, 'name': ReconfigVM_Task, 'duration_secs': 0.781784} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.229249] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 4b113b46-1da2-49ad-a0bb-c01b56c70f00/4b113b46-1da2-49ad-a0bb-c01b56c70f00.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.229995] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2945f41-4d8c-45df-ba32-56aaa4e4d98a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.236291] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 849.236291] env[61911]: value = "task-1251114" [ 849.236291] env[61911]: _type = "Task" [ 849.236291] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.247502] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251114, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.256962] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.278884] env[61911]: DEBUG oslo_concurrency.lockutils [req-4fdf9416-26d7-4a89-9a7c-43c231caecd1 req-d77703b0-79a0-4744-bacd-2cb5e4716fdd service nova] Releasing lock "refresh_cache-9230cd85-960d-427c-b69c-d19ac2bcfe35" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.392156] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.937s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.394954] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.690s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.396822] env[61911]: INFO nova.compute.claims [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.425736] env[61911]: INFO nova.scheduler.client.report [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Deleted allocations for instance 43d18895-202c-4048-9435-b3484ffd4c07 [ 849.514260] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251112, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.591977] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "refresh_cache-63d54f69-3d8f-45e9-b317-9d484af954b1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.592430] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "refresh_cache-63d54f69-3d8f-45e9-b317-9d484af954b1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.592430] env[61911]: DEBUG nova.network.neutron [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.661134] env[61911]: DEBUG oslo_vmware.api [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251110, 'name': ReconfigVM_Task, 'duration_secs': 1.086518} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.661501] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269594', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'name': 'volume-4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '28159e75-9fe9-44c7-b5c9-534902cecbef', 'attached_at': '', 'detached_at': '', 'volume_id': '4c9c79e1-b161-4bda-8c98-321d0adcdd98', 'serial': '4c9c79e1-b161-4bda-8c98-321d0adcdd98'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 849.673839] env[61911]: DEBUG nova.compute.manager [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Received event network-changed-28133c58-2aa2-41ae-9c8b-584659387203 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 849.674191] env[61911]: DEBUG nova.compute.manager [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Refreshing instance network info cache due to event network-changed-28133c58-2aa2-41ae-9c8b-584659387203. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 849.674306] env[61911]: DEBUG oslo_concurrency.lockutils [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] Acquiring lock "refresh_cache-63d54f69-3d8f-45e9-b317-9d484af954b1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.715605] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251113, 'name': Rename_Task, 'duration_secs': 0.168487} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.716403] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.716496] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2224dc0-ce9e-4efc-8b8b-0f2c3096ce65 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.722962] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 849.722962] env[61911]: value = "task-1251115" [ 849.722962] env[61911]: _type = "Task" [ 849.722962] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.733834] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.747923] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251114, 'name': Rename_Task, 'duration_secs': 0.263283} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.752317] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.752678] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72312f0d-d581-46e4-a824-9038bfaa7fd9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.759918] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.761267] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 849.761267] env[61911]: value = "task-1251116" [ 849.761267] env[61911]: _type = "Task" [ 849.761267] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.769362] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.934588] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4e59b798-effb-43df-af63-0e1c963b1c79 tempest-InstanceActionsTestJSON-771963648 tempest-InstanceActionsTestJSON-771963648-project-member] Lock "43d18895-202c-4048-9435-b3484ffd4c07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.961s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.014065] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251112, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.75047} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.014359] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. [ 850.015220] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711d333b-d6fe-4a7e-9e50-5fcbd2ee7b7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.041025] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.041368] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9eb9db84-4d83-41ca-94f5-6126f44443a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.060570] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 850.060570] env[61911]: value = "task-1251117" [ 850.060570] env[61911]: _type = "Task" [ 850.060570] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.068321] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.141127] env[61911]: DEBUG nova.network.neutron [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.219601] env[61911]: DEBUG nova.objects.instance [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lazy-loading 'flavor' on Instance uuid 28159e75-9fe9-44c7-b5c9-534902cecbef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.235538] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251115, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.260266] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.270972] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251116, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.394101] env[61911]: DEBUG nova.network.neutron [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Updating instance_info_cache with network_info: [{"id": "28133c58-2aa2-41ae-9c8b-584659387203", "address": "fa:16:3e:a1:89:3e", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28133c58-2a", "ovs_interfaceid": "28133c58-2aa2-41ae-9c8b-584659387203", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.577034] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251117, 'name': ReconfigVM_Task, 'duration_secs': 0.472414} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.577455] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Reconfigured VM instance instance-0000003d to attach disk [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.578657] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6064eaf1-c265-4526-8e94-fd9942cc0df0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.610920] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d3e4a8d-f086-48a7-96de-084fd358ec7e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.630846] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 850.630846] env[61911]: value = "task-1251118" [ 850.630846] env[61911]: _type = "Task" [ 850.630846] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.639395] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251118, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.733106] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b163ea-1e11-42c5-95aa-ea8a3bf3341c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.740071] env[61911]: DEBUG oslo_vmware.api [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251115, 'name': PowerOnVM_Task, 'duration_secs': 0.719325} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.740831] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.741280] env[61911]: INFO nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Took 11.52 seconds to spawn the instance on the hypervisor. [ 850.741570] env[61911]: DEBUG nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 850.742483] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b68850a-e508-4bbf-935e-1e664973b847 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.748419] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afee76c-a95f-40ab-9890-42671b174d0e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.806337] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.811238] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4822e4da-cf54-41da-9054-1720984397f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.819148] env[61911]: DEBUG oslo_vmware.api [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251116, 'name': PowerOnVM_Task, 'duration_secs': 0.91804} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.821334] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.821550] env[61911]: INFO nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Took 10.12 seconds to spawn the instance on the hypervisor. [ 850.821728] env[61911]: DEBUG nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 850.822543] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7c82c2-ca0c-4ae4-bcb4-a317f11d6a83 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.825764] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070e5fa1-9fb5-482b-8438-a6d5eac92b40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.840527] env[61911]: DEBUG nova.compute.provider_tree [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.896692] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "refresh_cache-63d54f69-3d8f-45e9-b317-9d484af954b1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.897036] env[61911]: DEBUG nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Instance network_info: |[{"id": "28133c58-2aa2-41ae-9c8b-584659387203", "address": "fa:16:3e:a1:89:3e", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28133c58-2a", "ovs_interfaceid": "28133c58-2aa2-41ae-9c8b-584659387203", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 850.897358] env[61911]: DEBUG oslo_concurrency.lockutils [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] Acquired lock "refresh_cache-63d54f69-3d8f-45e9-b317-9d484af954b1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.897565] env[61911]: DEBUG nova.network.neutron [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Refreshing network info cache for port 28133c58-2aa2-41ae-9c8b-584659387203 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.899116] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:89:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28133c58-2aa2-41ae-9c8b-584659387203', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.906956] env[61911]: DEBUG oslo.service.loopingcall [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.907888] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.908140] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c48471d-087f-431a-ada5-1d78b0a52c2d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.927683] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.927683] env[61911]: value = "task-1251119" [ 850.927683] env[61911]: _type = "Task" [ 850.927683] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.936531] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251119, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.140422] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251118, 'name': ReconfigVM_Task, 'duration_secs': 0.471504} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.140711] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.140998] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8b2ed11-a72c-4ea0-a1f1-23ab6bb07413 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.147272] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 851.147272] env[61911]: value = "task-1251120" [ 851.147272] env[61911]: _type = "Task" [ 851.147272] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.162341] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251120, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.236796] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fe210467-06b5-4ec5-9c10-98740ee26f4e tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.800s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.272482] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.274515] env[61911]: INFO nova.compute.manager [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Took 33.87 seconds to build instance. [ 851.347118] env[61911]: DEBUG nova.scheduler.client.report [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 851.359060] env[61911]: INFO nova.compute.manager [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Took 31.34 seconds to build instance. [ 851.437482] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251119, 'name': CreateVM_Task, 'duration_secs': 0.460371} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.437762] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.438796] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.439115] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.439499] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.440118] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-161eaae8-49d5-4881-9ac7-3d3d2b29b92c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.445083] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 851.445083] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5217bf49-a26b-89ec-e6d6-8c32e7c1fb74" [ 851.445083] env[61911]: _type = "Task" [ 851.445083] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.454374] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5217bf49-a26b-89ec-e6d6-8c32e7c1fb74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.664988] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251120, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.712798] env[61911]: DEBUG nova.network.neutron [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Updated VIF entry in instance network info cache for port 28133c58-2aa2-41ae-9c8b-584659387203. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.714442] env[61911]: DEBUG nova.network.neutron [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Updating instance_info_cache with network_info: [{"id": "28133c58-2aa2-41ae-9c8b-584659387203", "address": "fa:16:3e:a1:89:3e", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28133c58-2a", "ovs_interfaceid": "28133c58-2aa2-41ae-9c8b-584659387203", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.765206] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251108, 'name': CreateVM_Task, 'duration_secs': 3.794968} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.765417] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.766225] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.776862] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d52136fd-3d52-4220-8ac5-f9989a840ef2 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "6e100d92-eb95-4b9b-b236-34ff94051811" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.910s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.856019] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.856714] env[61911]: DEBUG nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 851.859445] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.226s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.860937] env[61911]: INFO nova.compute.claims [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.863876] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d829ab9d-9474-4cd0-87b5-802e6146bb31 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.853s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.875869] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "6e100d92-eb95-4b9b-b236-34ff94051811" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.875869] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "6e100d92-eb95-4b9b-b236-34ff94051811" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.876108] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "6e100d92-eb95-4b9b-b236-34ff94051811-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.876212] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "6e100d92-eb95-4b9b-b236-34ff94051811-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.876464] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "6e100d92-eb95-4b9b-b236-34ff94051811-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.878700] env[61911]: INFO nova.compute.manager [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Terminating instance [ 851.957845] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5217bf49-a26b-89ec-e6d6-8c32e7c1fb74, 'name': SearchDatastore_Task, 'duration_secs': 0.010466} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.958197] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.958504] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.958766] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.958921] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.959120] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.959408] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.959727] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.959981] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b1614a6-033e-4173-9296-55f7ddcedf86 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.961903] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b220c47-1a7c-4f09-8563-4a2a04cc3cc1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.967959] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 851.967959] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aac190-a1c7-557c-e74c-2e8055ae7938" [ 851.967959] env[61911]: _type = "Task" [ 851.967959] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.972560] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.972743] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.973737] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c714417-7cc6-4f22-be02-89d3b7a4a64a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.979157] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aac190-a1c7-557c-e74c-2e8055ae7938, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.982546] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 851.982546] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52006611-3cef-a043-23eb-15b81b505749" [ 851.982546] env[61911]: _type = "Task" [ 851.982546] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.989809] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52006611-3cef-a043-23eb-15b81b505749, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.158532] env[61911]: DEBUG oslo_vmware.api [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251120, 'name': PowerOnVM_Task, 'duration_secs': 0.678683} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.158821] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.164367] env[61911]: DEBUG nova.compute.manager [None req-44d06449-79ea-4734-94fc-b34e54e18d42 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 852.165508] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f6b7a8-c093-4ca0-b403-a63441a2ceea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.218162] env[61911]: DEBUG oslo_concurrency.lockutils [req-1cc5119a-75b6-4919-b29d-f8064ab4256b req-664e80e5-7517-4e9a-a5b4-32422c1392b1 service nova] Releasing lock "refresh_cache-63d54f69-3d8f-45e9-b317-9d484af954b1" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.366658] env[61911]: DEBUG nova.compute.utils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.370666] env[61911]: DEBUG nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 852.370666] env[61911]: DEBUG nova.network.neutron [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 852.382578] env[61911]: DEBUG nova.compute.manager [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 852.382871] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 852.383778] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5ac722-c056-4a10-9fd5-02fe072a6854 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.394278] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.394510] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff946e7f-ba3e-4f5c-835d-41fa7fa62e52 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.402982] env[61911]: DEBUG oslo_vmware.api [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 852.402982] env[61911]: value = "task-1251121" [ 852.402982] env[61911]: _type = "Task" [ 852.402982] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.412013] env[61911]: DEBUG oslo_vmware.api [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.452262] env[61911]: DEBUG nova.policy [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d0056b50ef7429ea460ff4c1156d132', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf7a7083777a43a5bb9fc04584dcdd06', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.478845] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aac190-a1c7-557c-e74c-2e8055ae7938, 'name': SearchDatastore_Task, 'duration_secs': 0.016776} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.479273] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.479494] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.479719] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.491655] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52006611-3cef-a043-23eb-15b81b505749, 'name': SearchDatastore_Task, 'duration_secs': 0.021006} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.492486] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d067a6c4-685e-4b63-9315-caf2ff6ded4e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.497743] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 852.497743] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b1f36e-eece-048d-8492-b3f276feb58b" [ 852.497743] env[61911]: _type = "Task" [ 852.497743] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.506270] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b1f36e-eece-048d-8492-b3f276feb58b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.870765] env[61911]: DEBUG nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 852.912837] env[61911]: DEBUG oslo_vmware.api [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251121, 'name': PowerOffVM_Task, 'duration_secs': 0.263732} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.917019] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.917350] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 852.917802] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3060200-bb37-4b84-ab24-061ae8bc4725 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.937816] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "b6b16408-7633-4f23-8f57-1193b1fb0384" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.939045] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.972127] env[61911]: DEBUG nova.network.neutron [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Successfully created port: 8548c265-98c0-4ef3-80a4-3e6d5b65516b {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.988133] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 852.988380] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 852.988618] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Deleting the datastore file [datastore2] 6e100d92-eb95-4b9b-b236-34ff94051811 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.991701] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-998dff97-32ee-42b5-b729-3e504d3c34b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.998072] env[61911]: DEBUG oslo_vmware.api [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for the task: (returnval){ [ 852.998072] env[61911]: value = "task-1251123" [ 852.998072] env[61911]: _type = "Task" [ 852.998072] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.014357] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b1f36e-eece-048d-8492-b3f276feb58b, 'name': SearchDatastore_Task, 'duration_secs': 0.011752} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.014597] env[61911]: DEBUG oslo_vmware.api [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.014848] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.015126] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 63d54f69-3d8f-45e9-b317-9d484af954b1/63d54f69-3d8f-45e9-b317-9d484af954b1.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.015411] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.015603] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.015815] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1cc4336-6634-471d-bd7d-c271015780c2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.017753] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ffaf1e4-0dc6-43f9-ab09-7bf0d98b64ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.026918] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.027191] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.027414] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.027604] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.027793] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.030215] env[61911]: INFO nova.compute.manager [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Terminating instance [ 853.036409] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 853.036409] env[61911]: value = "task-1251124" [ 853.036409] env[61911]: _type = "Task" [ 853.036409] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.040467] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.041198] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.042144] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e314d4a8-d706-4571-aba2-04af184f7c2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.053226] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.056623] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 853.056623] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523dfa93-52fd-9c49-ccb5-70b2a09bb8a8" [ 853.056623] env[61911]: _type = "Task" [ 853.056623] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.064407] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523dfa93-52fd-9c49-ccb5-70b2a09bb8a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.222417] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef2204b-feb7-49c0-93ce-45b30af25095 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.229798] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6084c6a-cf80-4ab5-9303-fd93d8088040 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.260844] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c77d23-1980-40cf-93f8-a8a184cf16bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.270515] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800ea738-040f-4e2c-94d0-9a7356b7acc6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.287255] env[61911]: DEBUG nova.compute.provider_tree [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.442636] env[61911]: DEBUG nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 853.509932] env[61911]: DEBUG oslo_vmware.api [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Task: {'id': task-1251123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217839} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.510332] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.510527] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 853.510711] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 853.510882] env[61911]: INFO nova.compute.manager [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Took 1.13 seconds to destroy the instance on the hypervisor. [ 853.511196] env[61911]: DEBUG oslo.service.loopingcall [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.511718] env[61911]: DEBUG nova.compute.manager [-] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 853.512012] env[61911]: DEBUG nova.network.neutron [-] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 853.535874] env[61911]: DEBUG nova.compute.manager [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 853.536128] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.537079] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e150319c-3be1-4389-aa8f-37fec9ec6a09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.549876] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251124, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.552120] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.552367] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4586e36c-bd27-4b16-baa9-fe5a5ab28dd3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.561212] env[61911]: DEBUG oslo_vmware.api [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 853.561212] env[61911]: value = "task-1251125" [ 853.561212] env[61911]: _type = "Task" [ 853.561212] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.568413] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523dfa93-52fd-9c49-ccb5-70b2a09bb8a8, 'name': SearchDatastore_Task, 'duration_secs': 0.010357} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.569764] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96812700-2c93-4a0d-a00d-6092b265e652 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.575477] env[61911]: DEBUG oslo_vmware.api [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.578487] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 853.578487] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5294055a-35c8-807f-ca6d-1370be303da4" [ 853.578487] env[61911]: _type = "Task" [ 853.578487] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.587175] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5294055a-35c8-807f-ca6d-1370be303da4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.791109] env[61911]: DEBUG nova.scheduler.client.report [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 853.889024] env[61911]: DEBUG nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 853.928898] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 853.929564] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.929564] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 853.929909] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.929959] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 853.930124] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 853.930445] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 853.930629] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 853.930833] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 853.931023] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 853.931219] env[61911]: DEBUG nova.virt.hardware [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 853.932492] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d38e29-1aea-4f35-8089-7730ef2cf849 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.943032] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e777e80f-8f37-403f-82ad-5459c6e1748d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.978604] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.055359] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251124, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53927} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.055359] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 63d54f69-3d8f-45e9-b317-9d484af954b1/63d54f69-3d8f-45e9-b317-9d484af954b1.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.055508] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.055765] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-77264199-d50b-456c-81a6-2f57bc3be02e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.063364] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 854.063364] env[61911]: value = "task-1251126" [ 854.063364] env[61911]: _type = "Task" [ 854.063364] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.081019] env[61911]: DEBUG oslo_vmware.api [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251125, 'name': PowerOffVM_Task, 'duration_secs': 0.24386} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.081318] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.081697] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.082198] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251126, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.082588] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54b760b0-91b6-4e03-87e5-bf5b50e49516 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.094329] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5294055a-35c8-807f-ca6d-1370be303da4, 'name': SearchDatastore_Task, 'duration_secs': 0.012378} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.094658] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.095071] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.095507] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-537242f1-6058-428e-ade8-1be47d01c1d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.104242] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 854.104242] env[61911]: value = "task-1251128" [ 854.104242] env[61911]: _type = "Task" [ 854.104242] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.113018] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251128, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.146173] env[61911]: DEBUG nova.compute.manager [req-af5debf9-e872-4c93-91b9-b48093af47d1 req-fe277721-bd2a-4fb1-b129-388022032100 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Received event network-vif-deleted-bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 854.148026] env[61911]: INFO nova.compute.manager [req-af5debf9-e872-4c93-91b9-b48093af47d1 req-fe277721-bd2a-4fb1-b129-388022032100 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Neutron deleted interface bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8; detaching it from the instance and deleting it from the info cache [ 854.148732] env[61911]: DEBUG nova.network.neutron [req-af5debf9-e872-4c93-91b9-b48093af47d1 req-fe277721-bd2a-4fb1-b129-388022032100 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.152848] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.153225] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.153555] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleting the datastore file [datastore2] 4b113b46-1da2-49ad-a0bb-c01b56c70f00 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.154450] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fc1d9c9-2469-4f01-a6d2-716127adfb7b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.168029] env[61911]: DEBUG oslo_vmware.api [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 854.168029] env[61911]: value = "task-1251129" [ 854.168029] env[61911]: _type = "Task" [ 854.168029] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.178928] env[61911]: DEBUG oslo_vmware.api [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.296754] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.297332] env[61911]: DEBUG nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 854.300113] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.301s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.300344] env[61911]: DEBUG nova.objects.instance [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lazy-loading 'resources' on Instance uuid 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.423800] env[61911]: INFO nova.compute.manager [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Unrescuing [ 854.423800] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.423800] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.423800] env[61911]: DEBUG nova.network.neutron [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.424439] env[61911]: DEBUG nova.network.neutron [-] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.551735] env[61911]: DEBUG nova.network.neutron [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Successfully updated port: 8548c265-98c0-4ef3-80a4-3e6d5b65516b {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.581092] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251126, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06649} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.581480] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.582370] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed88fd73-554e-491f-9780-80bc15a65ed8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.605266] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 63d54f69-3d8f-45e9-b317-9d484af954b1/63d54f69-3d8f-45e9-b317-9d484af954b1.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.606439] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed6a87b8-f207-4852-80ce-83a201565d9d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.630142] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251128, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.631845] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 854.631845] env[61911]: value = "task-1251130" [ 854.631845] env[61911]: _type = "Task" [ 854.631845] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.642721] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251130, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.652724] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5ee9a68-120b-4549-b9ee-ba70ba04a8fb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.662779] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752f940c-cd1f-4ac0-9675-922fb668b5b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.682309] env[61911]: DEBUG oslo_vmware.api [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.41548} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.682594] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.682783] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 854.682966] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 854.683162] env[61911]: INFO nova.compute.manager [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Took 1.15 seconds to destroy the instance on the hypervisor. [ 854.683413] env[61911]: DEBUG oslo.service.loopingcall [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.683619] env[61911]: DEBUG nova.compute.manager [-] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 854.683708] env[61911]: DEBUG nova.network.neutron [-] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.698443] env[61911]: DEBUG nova.compute.manager [req-af5debf9-e872-4c93-91b9-b48093af47d1 req-fe277721-bd2a-4fb1-b129-388022032100 service nova] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Detach interface failed, port_id=bdaa8bb9-f6ce-4725-ba96-8629e1f88cc8, reason: Instance 6e100d92-eb95-4b9b-b236-34ff94051811 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 854.803528] env[61911]: DEBUG nova.compute.utils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.808031] env[61911]: DEBUG nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 854.808240] env[61911]: DEBUG nova.network.neutron [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 854.928678] env[61911]: INFO nova.compute.manager [-] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Took 1.42 seconds to deallocate network for instance. [ 854.961964] env[61911]: DEBUG nova.policy [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ee96b52414148b7bff1d067595848b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f40f512384d4fedb75bfc53f3195243', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.054919] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "refresh_cache-f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.055893] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "refresh_cache-f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.055893] env[61911]: DEBUG nova.network.neutron [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.115996] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251128, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.736113} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.116987] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.116987] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.117326] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2eb50f8f-dbdc-4c84-95f6-469dab027389 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.126582] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 855.126582] env[61911]: value = "task-1251131" [ 855.126582] env[61911]: _type = "Task" [ 855.126582] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.135391] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251131, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.147147] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251130, 'name': ReconfigVM_Task, 'duration_secs': 0.440795} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.147147] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 63d54f69-3d8f-45e9-b317-9d484af954b1/63d54f69-3d8f-45e9-b317-9d484af954b1.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.147854] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3710676-5f77-4053-938a-378556249e71 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.156853] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 855.156853] env[61911]: value = "task-1251132" [ 855.156853] env[61911]: _type = "Task" [ 855.156853] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.175582] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251132, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.196044] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3c9eb8-7e68-4ce8-bc8f-213ee13f642b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.215758] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b12f541-85a3-4093-9f4b-fa4f5ba33168 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.251246] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7979d96e-d06a-45a1-aabb-c706fdcf09af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.262247] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499642ad-b5c2-475c-84b5-fc0f4fde2a73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.280447] env[61911]: DEBUG nova.compute.provider_tree [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.310260] env[61911]: DEBUG nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 855.438176] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.470274] env[61911]: DEBUG nova.network.neutron [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Updating instance_info_cache with network_info: [{"id": "15e642fd-30d2-41c2-98f7-0d19e729041e", "address": "fa:16:3e:8e:99:bd", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e642fd-30", "ovs_interfaceid": "15e642fd-30d2-41c2-98f7-0d19e729041e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.551522] env[61911]: DEBUG nova.network.neutron [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Successfully created port: de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.591114] env[61911]: DEBUG nova.network.neutron [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.645394] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251131, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066383} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.645690] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.646768] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d456c6-f961-43e1-99d8-837675de5e60 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.673204] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.677048] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8d20108-9859-4bdd-a2d9-2e0fb01e7bf0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.699581] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251132, 'name': Rename_Task, 'duration_secs': 0.146276} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.700837] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.701173] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 855.701173] env[61911]: value = "task-1251133" [ 855.701173] env[61911]: _type = "Task" [ 855.701173] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.701376] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-376f13c6-9467-4c62-bef5-c94eb00396da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.711907] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251133, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.713108] env[61911]: DEBUG nova.network.neutron [-] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.714260] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 855.714260] env[61911]: value = "task-1251134" [ 855.714260] env[61911]: _type = "Task" [ 855.714260] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.723236] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251134, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.783838] env[61911]: DEBUG nova.scheduler.client.report [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.792752] env[61911]: DEBUG nova.network.neutron [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Updating instance_info_cache with network_info: [{"id": "8548c265-98c0-4ef3-80a4-3e6d5b65516b", "address": "fa:16:3e:77:8b:4e", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8548c265-98", "ovs_interfaceid": "8548c265-98c0-4ef3-80a4-3e6d5b65516b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.974649] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-bafd58bc-8ae7-49a4-a039-31d328699010" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.975414] env[61911]: DEBUG nova.objects.instance [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lazy-loading 'flavor' on Instance uuid bafd58bc-8ae7-49a4-a039-31d328699010 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.215562] env[61911]: INFO nova.compute.manager [-] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Took 1.53 seconds to deallocate network for instance. [ 856.215998] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251133, 'name': ReconfigVM_Task, 'duration_secs': 0.273389} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.221045] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.224731] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1b0f5e9-c528-4dcf-86a9-1169b94b029f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.231457] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251134, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.232450] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 856.232450] env[61911]: value = "task-1251135" [ 856.232450] env[61911]: _type = "Task" [ 856.232450] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.243731] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251135, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.292418] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.296937] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.976s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.297281] env[61911]: DEBUG nova.objects.instance [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lazy-loading 'resources' on Instance uuid 65d310c6-d438-4c3d-bd44-f3b51123fe93 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.302619] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "refresh_cache-f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.302751] env[61911]: DEBUG nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Instance network_info: |[{"id": "8548c265-98c0-4ef3-80a4-3e6d5b65516b", "address": "fa:16:3e:77:8b:4e", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8548c265-98", "ovs_interfaceid": "8548c265-98c0-4ef3-80a4-3e6d5b65516b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 856.303179] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:8b:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8548c265-98c0-4ef3-80a4-3e6d5b65516b', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.316623] env[61911]: DEBUG oslo.service.loopingcall [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.320548] env[61911]: INFO nova.scheduler.client.report [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Deleted allocations for instance 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf [ 856.322597] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.324086] env[61911]: DEBUG nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 856.328936] env[61911]: DEBUG nova.compute.manager [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Received event network-vif-plugged-8548c265-98c0-4ef3-80a4-3e6d5b65516b {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 856.329045] env[61911]: DEBUG oslo_concurrency.lockutils [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] Acquiring lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.329310] env[61911]: DEBUG oslo_concurrency.lockutils [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.329563] env[61911]: DEBUG oslo_concurrency.lockutils [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.329806] env[61911]: DEBUG nova.compute.manager [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] No waiting events found dispatching network-vif-plugged-8548c265-98c0-4ef3-80a4-3e6d5b65516b {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 856.330553] env[61911]: WARNING nova.compute.manager [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Received unexpected event network-vif-plugged-8548c265-98c0-4ef3-80a4-3e6d5b65516b for instance with vm_state building and task_state spawning. [ 856.330553] env[61911]: DEBUG nova.compute.manager [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Received event network-changed-8548c265-98c0-4ef3-80a4-3e6d5b65516b {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 856.330553] env[61911]: DEBUG nova.compute.manager [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Refreshing instance network info cache due to event network-changed-8548c265-98c0-4ef3-80a4-3e6d5b65516b. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 856.330970] env[61911]: DEBUG oslo_concurrency.lockutils [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] Acquiring lock "refresh_cache-f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.330970] env[61911]: DEBUG oslo_concurrency.lockutils [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] Acquired lock "refresh_cache-f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.331196] env[61911]: DEBUG nova.network.neutron [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Refreshing network info cache for port 8548c265-98c0-4ef3-80a4-3e6d5b65516b {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 856.337074] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb09486b-e6ac-4a65-89b8-53582963050c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.363760] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.363760] env[61911]: value = "task-1251136" [ 856.363760] env[61911]: _type = "Task" [ 856.363760] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.370070] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 856.370391] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.370803] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 856.370803] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.371013] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 856.371265] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 856.371590] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 856.372119] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 856.372119] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 856.372432] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 856.372616] env[61911]: DEBUG nova.virt.hardware [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 856.374612] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263b4f11-8e7d-4851-9ce5-d073407b3375 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.384200] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251136, 'name': CreateVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.388999] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d423f4-a13c-4e6c-9e37-03b2fac90be8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.482954] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c3987f-6789-4d06-ab48-cc726ee45dde {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.506885] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 856.507323] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-776fe876-688f-4b66-b6f4-b6e1e2c5b030 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.518088] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 856.518088] env[61911]: value = "task-1251137" [ 856.518088] env[61911]: _type = "Task" [ 856.518088] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.526752] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.725690] env[61911]: DEBUG oslo_vmware.api [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251134, 'name': PowerOnVM_Task, 'duration_secs': 0.751959} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.727209] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.727825] env[61911]: INFO nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Took 9.20 seconds to spawn the instance on the hypervisor. [ 856.727825] env[61911]: DEBUG nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 856.728724] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.731942] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b2b7f6-9a7a-44a6-9cad-2caa33767060 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.735085] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "5b7d09e1-ee14-4001-8d73-14e763402670" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.735316] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.758402] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251135, 'name': Rename_Task, 'duration_secs': 0.14543} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.759027] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.759027] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5338fc12-ab65-469c-b3f1-8791cbeb3e16 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.765664] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 856.765664] env[61911]: value = "task-1251138" [ 856.765664] env[61911]: _type = "Task" [ 856.765664] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.774160] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.863156] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cd8b72e8-8cde-4960-9df8-e9e619b9767f tempest-ServerTagsTestJSON-1433128332 tempest-ServerTagsTestJSON-1433128332-project-member] Lock "0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.388s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.875338] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251136, 'name': CreateVM_Task, 'duration_secs': 0.411164} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.876471] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.877211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.877620] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.877969] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.878778] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef61ea63-ada5-49ad-9e04-02bd9e6c9a0e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.890154] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 856.890154] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c8315c-0a6c-2bf6-b29a-a3e4625a5314" [ 856.890154] env[61911]: _type = "Task" [ 856.890154] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.899712] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c8315c-0a6c-2bf6-b29a-a3e4625a5314, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.031751] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251137, 'name': PowerOffVM_Task, 'duration_secs': 0.440751} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.031751] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.037692] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Reconfiguring VM instance instance-0000003d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 857.038300] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6a5abbb-77ce-4db8-b733-cbb12b00b53c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.062040] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 857.062040] env[61911]: value = "task-1251139" [ 857.062040] env[61911]: _type = "Task" [ 857.062040] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.069089] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251139, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.093133] env[61911]: DEBUG nova.network.neutron [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Updated VIF entry in instance network info cache for port 8548c265-98c0-4ef3-80a4-3e6d5b65516b. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.093488] env[61911]: DEBUG nova.network.neutron [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Updating instance_info_cache with network_info: [{"id": "8548c265-98c0-4ef3-80a4-3e6d5b65516b", "address": "fa:16:3e:77:8b:4e", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8548c265-98", "ovs_interfaceid": "8548c265-98c0-4ef3-80a4-3e6d5b65516b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.132997] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71416a8-5b11-4231-a650-b36734cba001 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.140812] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be319094-0cf2-45cb-ac54-574ee16103ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.173256] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d1dc3a-d67f-4533-91ee-3c987a3cb8ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.181543] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88ea6ff-75b7-4733-878d-f8bcd4e25917 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.200289] env[61911]: DEBUG nova.compute.provider_tree [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.244349] env[61911]: DEBUG nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 857.260039] env[61911]: INFO nova.compute.manager [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Took 31.12 seconds to build instance. [ 857.276594] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251138, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.400348] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c8315c-0a6c-2bf6-b29a-a3e4625a5314, 'name': SearchDatastore_Task, 'duration_secs': 0.013488} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.400706] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.400955] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.401217] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.401371] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.401557] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.401823] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e2a73e3-be3f-422d-b003-fe8af2833acd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.410195] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.410393] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.411140] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ebe929f-0fad-4f15-a585-cde8e7194fd9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.416364] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 857.416364] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526602e2-c799-cfe2-e591-71de22f941bf" [ 857.416364] env[61911]: _type = "Task" [ 857.416364] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.424505] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526602e2-c799-cfe2-e591-71de22f941bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.571085] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251139, 'name': ReconfigVM_Task, 'duration_secs': 0.349999} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.571441] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Reconfigured VM instance instance-0000003d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 857.571622] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.571863] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d6217c4-71da-4ed6-8039-9db07b973805 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.578587] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 857.578587] env[61911]: value = "task-1251140" [ 857.578587] env[61911]: _type = "Task" [ 857.578587] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.587347] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.598152] env[61911]: DEBUG oslo_concurrency.lockutils [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] Releasing lock "refresh_cache-f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.598470] env[61911]: DEBUG nova.compute.manager [req-d98e4be4-5771-42a4-ad19-869bd95349e0 req-ed029e29-3a19-4648-8af8-2fa60bc957fe service nova] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Received event network-vif-deleted-21db17c6-a0d9-48bc-96ed-b51ba77e6d40 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 857.700167] env[61911]: DEBUG nova.scheduler.client.report [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.761147] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7f2ca93e-e17a-42d8-a953-e5189e42a084 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.117s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.768650] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.778382] env[61911]: DEBUG oslo_vmware.api [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251138, 'name': PowerOnVM_Task, 'duration_secs': 0.58751} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.778825] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.779106] env[61911]: INFO nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Took 12.77 seconds to spawn the instance on the hypervisor. [ 857.779335] env[61911]: DEBUG nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 857.780150] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50cca350-fe06-4daa-96ef-0a08537e668b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.805526] env[61911]: DEBUG nova.network.neutron [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Successfully updated port: de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.930846] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526602e2-c799-cfe2-e591-71de22f941bf, 'name': SearchDatastore_Task, 'duration_secs': 0.018231} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.933024] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aae302f3-4e00-4a97-a1e5-086667dc3ae0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.940706] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 857.940706] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523fa8b6-ff7c-3303-4140-315c2f7e2183" [ 857.940706] env[61911]: _type = "Task" [ 857.940706] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.953026] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523fa8b6-ff7c-3303-4140-315c2f7e2183, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.089117] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251140, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.206978] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.910s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.207938] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.663s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.211277] env[61911]: INFO nova.compute.claims [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.227719] env[61911]: INFO nova.scheduler.client.report [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Deleted allocations for instance 65d310c6-d438-4c3d-bd44-f3b51123fe93 [ 858.302688] env[61911]: INFO nova.compute.manager [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Took 34.83 seconds to build instance. [ 858.312553] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.313324] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.314972] env[61911]: DEBUG nova.network.neutron [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.355249] env[61911]: DEBUG nova.compute.manager [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-vif-plugged-de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 858.355249] env[61911]: DEBUG oslo_concurrency.lockutils [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] Acquiring lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.355249] env[61911]: DEBUG oslo_concurrency.lockutils [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.355249] env[61911]: DEBUG oslo_concurrency.lockutils [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.355249] env[61911]: DEBUG nova.compute.manager [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] No waiting events found dispatching network-vif-plugged-de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 858.355249] env[61911]: WARNING nova.compute.manager [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received unexpected event network-vif-plugged-de067cc9-827d-4e43-a53e-6eda53bf9d1e for instance with vm_state building and task_state spawning. [ 858.355249] env[61911]: DEBUG nova.compute.manager [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 858.355249] env[61911]: DEBUG nova.compute.manager [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing instance network info cache due to event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 858.355958] env[61911]: DEBUG oslo_concurrency.lockutils [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.453145] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523fa8b6-ff7c-3303-4140-315c2f7e2183, 'name': SearchDatastore_Task, 'duration_secs': 0.015394} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.453145] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.453145] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] f0dcfbbf-98ab-4e9f-838f-d9601d2bb045/f0dcfbbf-98ab-4e9f-838f-d9601d2bb045.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.455382] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ead72bab-197e-47db-be76-b0fa3d0a2bb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.460825] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 858.460825] env[61911]: value = "task-1251141" [ 858.460825] env[61911]: _type = "Task" [ 858.460825] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.469143] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251141, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.597039] env[61911]: DEBUG oslo_vmware.api [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251140, 'name': PowerOnVM_Task, 'duration_secs': 0.561598} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.597039] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.597039] env[61911]: DEBUG nova.compute.manager [None req-7844228b-8e39-42dc-9078-dce0f25558f9 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 858.597039] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c745293-f826-44ef-aa8a-a4ae3241c2ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.738922] env[61911]: DEBUG oslo_concurrency.lockutils [None req-829b587d-7561-43be-a30a-3c16a24dbd20 tempest-ServersAaction247Test-1313704054 tempest-ServersAaction247Test-1313704054-project-member] Lock "65d310c6-d438-4c3d-bd44-f3b51123fe93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.684s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.805814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2373f26a-55fe-4ae0-b4ea-cd46444e55bd tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.411s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.868737] env[61911]: DEBUG nova.network.neutron [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.974025] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "63d54f69-3d8f-45e9-b317-9d484af954b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.974025] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.974417] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "63d54f69-3d8f-45e9-b317-9d484af954b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.974664] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.974847] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.977701] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251141, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.978404] env[61911]: INFO nova.compute.manager [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Terminating instance [ 859.178346] env[61911]: DEBUG nova.network.neutron [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.475416] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251141, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656467} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.475520] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] f0dcfbbf-98ab-4e9f-838f-d9601d2bb045/f0dcfbbf-98ab-4e9f-838f-d9601d2bb045.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.476028] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.476375] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62925c1a-334a-4671-b53b-146dad10f3ff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.484271] env[61911]: DEBUG nova.compute.manager [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 859.484486] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.487337] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac5350a-c8d2-46bd-a215-5d2dc622f63d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.491435] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 859.491435] env[61911]: value = "task-1251142" [ 859.491435] env[61911]: _type = "Task" [ 859.491435] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.497707] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.499017] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81622883-402d-4251-ad06-270ce1ab322f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.503924] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251142, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.511577] env[61911]: DEBUG oslo_vmware.api [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 859.511577] env[61911]: value = "task-1251143" [ 859.511577] env[61911]: _type = "Task" [ 859.511577] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.522216] env[61911]: DEBUG oslo_vmware.api [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.568610] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9138a86-ba3f-42a7-9664-f8004a657a36 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.577225] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f4dc63-0002-4cbc-b89c-3b5446db45aa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.613423] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f5a3ca-7742-4f2f-94b7-e9d708a160c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.621670] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc5d9b5-a09d-4068-9854-055203c15441 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.640947] env[61911]: DEBUG nova.compute.provider_tree [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.681838] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.682237] env[61911]: DEBUG nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Instance network_info: |[{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 859.682614] env[61911]: DEBUG oslo_concurrency.lockutils [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.682765] env[61911]: DEBUG nova.network.neutron [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.686049] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:05:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de067cc9-827d-4e43-a53e-6eda53bf9d1e', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.691992] env[61911]: DEBUG oslo.service.loopingcall [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.695131] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.695131] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd01fc54-a606-427a-9c4f-35093d82c64d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.714211] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.714211] env[61911]: value = "task-1251144" [ 859.714211] env[61911]: _type = "Task" [ 859.714211] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.722698] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251144, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.001522] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251142, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074452} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.001522] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.002359] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3792de89-b532-4ac8-a8cb-549e0f49c592 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.025132] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] f0dcfbbf-98ab-4e9f-838f-d9601d2bb045/f0dcfbbf-98ab-4e9f-838f-d9601d2bb045.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.028669] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b5ef227-7216-448a-a7ab-b516436a679d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.048410] env[61911]: DEBUG oslo_vmware.api [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251143, 'name': PowerOffVM_Task, 'duration_secs': 0.180194} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.049969] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.049969] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.050352] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 860.050352] env[61911]: value = "task-1251145" [ 860.050352] env[61911]: _type = "Task" [ 860.050352] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.050697] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ee276eb-6483-4a79-832f-45e67b9857bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.059833] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251145, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.119190] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.119318] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.119776] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleting the datastore file [datastore1] 63d54f69-3d8f-45e9-b317-9d484af954b1 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.119776] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2dd1aad-059c-4141-9d65-68e93de6d5a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.126249] env[61911]: DEBUG oslo_vmware.api [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 860.126249] env[61911]: value = "task-1251147" [ 860.126249] env[61911]: _type = "Task" [ 860.126249] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.134759] env[61911]: DEBUG oslo_vmware.api [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.145218] env[61911]: DEBUG nova.scheduler.client.report [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 860.227942] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251144, 'name': CreateVM_Task, 'duration_secs': 0.332501} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.228257] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.233020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.233020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.233020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.233020] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dd5149f-9cc8-4ae6-96b5-b38c55d2e560 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.236953] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 860.236953] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac7630-0cdc-009c-bbbd-4805dcf115a4" [ 860.236953] env[61911]: _type = "Task" [ 860.236953] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.253833] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac7630-0cdc-009c-bbbd-4805dcf115a4, 'name': SearchDatastore_Task, 'duration_secs': 0.009952} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.253833] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.253833] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.253833] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.254157] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.254337] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.254754] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee2b48c0-39e2-4659-92df-52c7abb700d7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.264156] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.264360] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.265157] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6df9dd3f-68e6-4178-ab28-ec2590d6d568 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.270352] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 860.270352] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac832a-bf53-0983-aebc-c2f900f0599f" [ 860.270352] env[61911]: _type = "Task" [ 860.270352] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.279884] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac832a-bf53-0983-aebc-c2f900f0599f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.313272] env[61911]: INFO nova.compute.manager [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Rebuilding instance [ 860.380864] env[61911]: DEBUG nova.compute.manager [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 860.382504] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628316dc-c81f-4d6f-9a9c-0be79d8b77c8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.560394] env[61911]: DEBUG nova.network.neutron [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updated VIF entry in instance network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 860.560860] env[61911]: DEBUG nova.network.neutron [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.567288] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251145, 'name': ReconfigVM_Task, 'duration_secs': 0.384338} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.567919] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Reconfigured VM instance instance-00000046 to attach disk [datastore1] f0dcfbbf-98ab-4e9f-838f-d9601d2bb045/f0dcfbbf-98ab-4e9f-838f-d9601d2bb045.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.568589] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3eac822-b94c-49e1-b01d-a4fef23cbd15 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.575907] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 860.575907] env[61911]: value = "task-1251148" [ 860.575907] env[61911]: _type = "Task" [ 860.575907] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.584532] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251148, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.636711] env[61911]: DEBUG oslo_vmware.api [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.235864} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.636981] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.637277] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.637501] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.637717] env[61911]: INFO nova.compute.manager [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Took 1.15 seconds to destroy the instance on the hypervisor. [ 860.638129] env[61911]: DEBUG oslo.service.loopingcall [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.638245] env[61911]: DEBUG nova.compute.manager [-] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 860.638334] env[61911]: DEBUG nova.network.neutron [-] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.650328] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.650834] env[61911]: DEBUG nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 860.653371] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.555s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.654834] env[61911]: INFO nova.compute.claims [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.784020] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac832a-bf53-0983-aebc-c2f900f0599f, 'name': SearchDatastore_Task, 'duration_secs': 0.033206} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.784020] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b05e8575-e6d1-40cb-ad8c-513f92bc8c5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.790017] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 860.790017] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52141a80-b024-6e19-6e7c-f6c1348f08c3" [ 860.790017] env[61911]: _type = "Task" [ 860.790017] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.796343] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52141a80-b024-6e19-6e7c-f6c1348f08c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.935331] env[61911]: DEBUG nova.compute.manager [req-97b3228a-7b7b-4ac1-b053-d04ea93e585d req-033691f0-c2a1-42c4-bb68-c41e02e42c90 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Received event network-vif-deleted-28133c58-2aa2-41ae-9c8b-584659387203 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 860.936047] env[61911]: INFO nova.compute.manager [req-97b3228a-7b7b-4ac1-b053-d04ea93e585d req-033691f0-c2a1-42c4-bb68-c41e02e42c90 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Neutron deleted interface 28133c58-2aa2-41ae-9c8b-584659387203; detaching it from the instance and deleting it from the info cache [ 860.936322] env[61911]: DEBUG nova.network.neutron [req-97b3228a-7b7b-4ac1-b053-d04ea93e585d req-033691f0-c2a1-42c4-bb68-c41e02e42c90 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.069103] env[61911]: DEBUG oslo_concurrency.lockutils [req-efb193f2-a9c9-4253-a8e8-80f210c242e6 req-6e5ebde8-1021-44df-acc3-4a808e110b0b service nova] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.089865] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251148, 'name': Rename_Task, 'duration_secs': 0.15336} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.090364] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.090862] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-204dd95d-b877-44fa-bf8c-56bdaaa049dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.096828] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 861.096828] env[61911]: value = "task-1251149" [ 861.096828] env[61911]: _type = "Task" [ 861.096828] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.106733] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.161029] env[61911]: DEBUG nova.compute.utils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.163271] env[61911]: DEBUG nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 861.163767] env[61911]: DEBUG nova.network.neutron [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 861.213889] env[61911]: DEBUG nova.policy [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8593a447cc994e1eb0d61a548ba98ba8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e38ba711d3e3431b9d258c91ee1ce6f1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.302951] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52141a80-b024-6e19-6e7c-f6c1348f08c3, 'name': SearchDatastore_Task, 'duration_secs': 0.010135} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.303394] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.304602] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] e3ad3416-423c-4687-aa35-2fb2ef67ec64/e3ad3416-423c-4687-aa35-2fb2ef67ec64.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.304602] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-95a6eacd-0182-4db7-8fcc-23bd0f36865a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.312735] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 861.312735] env[61911]: value = "task-1251150" [ 861.312735] env[61911]: _type = "Task" [ 861.312735] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.328530] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251150, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.403615] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.403946] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6031b311-8491-41d2-851f-d508c5fa0625 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.412752] env[61911]: DEBUG nova.network.neutron [-] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.417562] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 861.417562] env[61911]: value = "task-1251151" [ 861.417562] env[61911]: _type = "Task" [ 861.417562] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.428094] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251151, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.440314] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e28816fe-c9c6-4a07-835d-50d3afb78590 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.450167] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7df63d-b6fe-45b4-8dc6-7815cd0ff318 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.495504] env[61911]: DEBUG nova.compute.manager [req-97b3228a-7b7b-4ac1-b053-d04ea93e585d req-033691f0-c2a1-42c4-bb68-c41e02e42c90 service nova] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Detach interface failed, port_id=28133c58-2aa2-41ae-9c8b-584659387203, reason: Instance 63d54f69-3d8f-45e9-b317-9d484af954b1 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 861.609211] env[61911]: DEBUG oslo_vmware.api [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251149, 'name': PowerOnVM_Task, 'duration_secs': 0.497941} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.609625] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.609965] env[61911]: INFO nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Took 7.72 seconds to spawn the instance on the hypervisor. [ 861.610124] env[61911]: DEBUG nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 861.611208] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d5a2c0-3811-4462-9e91-bd7d45b51d2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.664622] env[61911]: DEBUG nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 861.831726] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251150, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.917113] env[61911]: INFO nova.compute.manager [-] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Took 1.28 seconds to deallocate network for instance. [ 861.933414] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251151, 'name': PowerOffVM_Task, 'duration_secs': 0.209537} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.933700] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 861.933933] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.934748] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4030bfbf-4053-4bce-9a62-e0b8bd18841e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.942458] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 861.942729] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf0174d5-82d5-426b-af17-e7e45f8ccaed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.040762] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74df32db-c856-4cd9-a634-6bb4ca138ec0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.048526] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4b7aff-4c4e-4f31-98fa-3ee58547c652 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.090173] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d0136f-e2a4-4403-97d7-cdb086c2094b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.098709] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca9709a1-e809-43c3-af69-f2c597f98eea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.103457] env[61911]: DEBUG nova.network.neutron [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Successfully created port: 54a15f3e-c3d1-45b1-ac13-19e74bed26c6 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.115581] env[61911]: DEBUG nova.compute.provider_tree [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.140376] env[61911]: INFO nova.compute.manager [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Took 31.45 seconds to build instance. [ 862.275816] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 862.275935] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 862.276163] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleting the datastore file [datastore1] 9230cd85-960d-427c-b69c-d19ac2bcfe35 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.276472] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa6787a9-f5ee-4c8b-99a4-8933946d45bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.287047] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 862.287047] env[61911]: value = "task-1251153" [ 862.287047] env[61911]: _type = "Task" [ 862.287047] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.293779] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251153, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.324762] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251150, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542706} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.325134] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] e3ad3416-423c-4687-aa35-2fb2ef67ec64/e3ad3416-423c-4687-aa35-2fb2ef67ec64.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.325368] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.325632] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3387d62-d736-4710-9ab1-718578e9b781 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.333918] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 862.333918] env[61911]: value = "task-1251154" [ 862.333918] env[61911]: _type = "Task" [ 862.333918] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.344799] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.430718] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.618580] env[61911]: DEBUG nova.scheduler.client.report [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 862.643233] env[61911]: DEBUG oslo_concurrency.lockutils [None req-24d3397f-79c1-456e-9555-8b4378dbfb28 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.769s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.675175] env[61911]: DEBUG nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 862.707203] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 862.707567] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.707767] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 862.708126] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.708374] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 862.708546] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 862.708769] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 862.709586] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 862.709586] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 862.709586] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 862.709586] env[61911]: DEBUG nova.virt.hardware [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 862.712843] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a295b1-8477-444a-9346-6eb19897d3b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.720011] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257263ce-4223-43c9-88fd-c91f124b378f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.798244] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251153, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.846760] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109828} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.847223] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.848073] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a488dd-de1a-4eab-95f2-e3d8acc8b16a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.877750] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] e3ad3416-423c-4687-aa35-2fb2ef67ec64/e3ad3416-423c-4687-aa35-2fb2ef67ec64.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.878167] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d27db1ba-f9b8-4dd5-9d73-b64a44001c72 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.900990] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 862.900990] env[61911]: value = "task-1251155" [ 862.900990] env[61911]: _type = "Task" [ 862.900990] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.915444] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251155, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.129650] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.130219] env[61911]: DEBUG nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 863.132772] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.154s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.134210] env[61911]: INFO nova.compute.claims [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.299927] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251153, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.623282} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.300352] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.300593] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 863.300762] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 863.414118] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251155, 'name': ReconfigVM_Task, 'duration_secs': 0.277672} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.414418] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Reconfigured VM instance instance-00000047 to attach disk [datastore2] e3ad3416-423c-4687-aa35-2fb2ef67ec64/e3ad3416-423c-4687-aa35-2fb2ef67ec64.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.415226] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c08952ba-7691-43d0-8daa-e3caef261e43 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.427267] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 863.427267] env[61911]: value = "task-1251156" [ 863.427267] env[61911]: _type = "Task" [ 863.427267] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.443947] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251156, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.499584] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "da6f7457-e603-415c-9578-024a7563d34f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.499787] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "da6f7457-e603-415c-9578-024a7563d34f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.638425] env[61911]: DEBUG nova.compute.utils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.643436] env[61911]: DEBUG nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 863.643619] env[61911]: DEBUG nova.network.neutron [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.690559] env[61911]: DEBUG nova.policy [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b89d8654369c48f4b39a1a0e8c926bdd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ada275c0bead46f8a9d295952cd930ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 863.870534] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.871080] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.943075] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251156, 'name': Rename_Task, 'duration_secs': 0.163478} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.943075] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 863.943075] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f3d294e-0096-49fd-bcb6-8d65b9909c7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.952591] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 863.952591] env[61911]: value = "task-1251157" [ 863.952591] env[61911]: _type = "Task" [ 863.952591] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.962052] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.982017] env[61911]: DEBUG nova.network.neutron [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Successfully created port: 0e16550a-783a-473b-9be5-cb569b373191 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.004679] env[61911]: DEBUG nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 864.015377] env[61911]: DEBUG nova.compute.manager [req-1564c54a-f8b5-46c0-af87-68fe193fb18a req-6a5ab7f1-349e-4c07-a569-9f39854326a7 service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Received event network-vif-plugged-54a15f3e-c3d1-45b1-ac13-19e74bed26c6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 864.019019] env[61911]: DEBUG oslo_concurrency.lockutils [req-1564c54a-f8b5-46c0-af87-68fe193fb18a req-6a5ab7f1-349e-4c07-a569-9f39854326a7 service nova] Acquiring lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.019019] env[61911]: DEBUG oslo_concurrency.lockutils [req-1564c54a-f8b5-46c0-af87-68fe193fb18a req-6a5ab7f1-349e-4c07-a569-9f39854326a7 service nova] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.019019] env[61911]: DEBUG oslo_concurrency.lockutils [req-1564c54a-f8b5-46c0-af87-68fe193fb18a req-6a5ab7f1-349e-4c07-a569-9f39854326a7 service nova] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.019019] env[61911]: DEBUG nova.compute.manager [req-1564c54a-f8b5-46c0-af87-68fe193fb18a req-6a5ab7f1-349e-4c07-a569-9f39854326a7 service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] No waiting events found dispatching network-vif-plugged-54a15f3e-c3d1-45b1-ac13-19e74bed26c6 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 864.019019] env[61911]: WARNING nova.compute.manager [req-1564c54a-f8b5-46c0-af87-68fe193fb18a req-6a5ab7f1-349e-4c07-a569-9f39854326a7 service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Received unexpected event network-vif-plugged-54a15f3e-c3d1-45b1-ac13-19e74bed26c6 for instance with vm_state building and task_state spawning. [ 864.142623] env[61911]: DEBUG nova.network.neutron [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Successfully updated port: 54a15f3e-c3d1-45b1-ac13-19e74bed26c6 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.152066] env[61911]: DEBUG nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 864.354443] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 864.354827] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.355051] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 864.355296] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.355494] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 864.355688] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 864.355948] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 864.356176] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 864.356404] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 864.356652] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 864.356885] env[61911]: DEBUG nova.virt.hardware [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 864.358340] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548131e2-fe9b-4243-b79b-144a38b81472 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.373144] env[61911]: DEBUG nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 864.377773] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fdafbc-7893-4f76-bfc1-9fc48e4a3a4f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.395059] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:bd:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47a25084-ef5a-4059-92b8-d93d72091857', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.404911] env[61911]: DEBUG oslo.service.loopingcall [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.408065] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.408527] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f226e5a-a2d0-4c8b-8ba7-1ebed1b64a7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.432690] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.432690] env[61911]: value = "task-1251158" [ 864.432690] env[61911]: _type = "Task" [ 864.432690] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.442770] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251158, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.462698] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251157, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.529518] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.559986] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d59db19-9927-4f95-ba78-8e4e054c05fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.568307] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ad7ea2-dfa7-45f5-9e80-c52501aa8189 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.599389] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73192e5-34eb-4969-b084-c0af7b771462 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.607555] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1973b9e9-64e5-4f36-a889-9ea15505ca0c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.624068] env[61911]: DEBUG nova.compute.provider_tree [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.645271] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.645409] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.645611] env[61911]: DEBUG nova.network.neutron [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 864.680135] env[61911]: DEBUG nova.compute.manager [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 864.680972] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4819a0cc-2dff-45d7-9e1c-ffe8c01f184d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.898617] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.942540] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251158, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.964385] env[61911]: DEBUG oslo_vmware.api [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251157, 'name': PowerOnVM_Task, 'duration_secs': 0.60338} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.964786] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.965041] env[61911]: INFO nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Took 8.64 seconds to spawn the instance on the hypervisor. [ 864.965236] env[61911]: DEBUG nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 864.966172] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa7b7b3-8443-4f41-8f4b-a46964a066d1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.127017] env[61911]: DEBUG nova.scheduler.client.report [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 865.162901] env[61911]: DEBUG nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 865.182118] env[61911]: DEBUG nova.network.neutron [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.186262] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 865.186530] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.186817] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 865.187090] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.187300] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 865.187494] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 865.187760] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 865.187984] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 865.188219] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 865.188437] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 865.188658] env[61911]: DEBUG nova.virt.hardware [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 865.189732] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0bd0787-2caa-48ea-a315-9223f055757c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.192967] env[61911]: INFO nova.compute.manager [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] instance snapshotting [ 865.195699] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2b6e11-44bd-470f-b04d-8de6d701290e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.203638] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb6a117-e27e-4edb-b13f-404114692274 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.221478] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b420b821-aa3b-4a6f-bebe-becda7729312 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.338077] env[61911]: DEBUG nova.network.neutron [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Updating instance_info_cache with network_info: [{"id": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "address": "fa:16:3e:3d:35:a7", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54a15f3e-c3", "ovs_interfaceid": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.443027] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251158, 'name': CreateVM_Task, 'duration_secs': 0.804967} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.443227] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.443864] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.444035] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.444364] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.444621] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1706a7e7-32fb-4457-aa10-79c35127be86 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.450113] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 865.450113] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529df7dc-1dbf-a9b2-f0b7-23dbfaf30669" [ 865.450113] env[61911]: _type = "Task" [ 865.450113] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.458851] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529df7dc-1dbf-a9b2-f0b7-23dbfaf30669, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.484068] env[61911]: INFO nova.compute.manager [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Took 32.86 seconds to build instance. [ 865.497841] env[61911]: DEBUG nova.network.neutron [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Successfully updated port: 0e16550a-783a-473b-9be5-cb569b373191 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.512687] env[61911]: DEBUG nova.compute.manager [req-1484908d-09fe-41c6-a876-2dce2bc9a448 req-23891d7c-971c-4345-b883-b0f678297774 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Received event network-vif-plugged-0e16550a-783a-473b-9be5-cb569b373191 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 865.512952] env[61911]: DEBUG oslo_concurrency.lockutils [req-1484908d-09fe-41c6-a876-2dce2bc9a448 req-23891d7c-971c-4345-b883-b0f678297774 service nova] Acquiring lock "1058f8d2-0979-4423-802d-52d91d5a2ff7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.513283] env[61911]: DEBUG oslo_concurrency.lockutils [req-1484908d-09fe-41c6-a876-2dce2bc9a448 req-23891d7c-971c-4345-b883-b0f678297774 service nova] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.513467] env[61911]: DEBUG oslo_concurrency.lockutils [req-1484908d-09fe-41c6-a876-2dce2bc9a448 req-23891d7c-971c-4345-b883-b0f678297774 service nova] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.513712] env[61911]: DEBUG nova.compute.manager [req-1484908d-09fe-41c6-a876-2dce2bc9a448 req-23891d7c-971c-4345-b883-b0f678297774 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] No waiting events found dispatching network-vif-plugged-0e16550a-783a-473b-9be5-cb569b373191 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 865.513904] env[61911]: WARNING nova.compute.manager [req-1484908d-09fe-41c6-a876-2dce2bc9a448 req-23891d7c-971c-4345-b883-b0f678297774 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Received unexpected event network-vif-plugged-0e16550a-783a-473b-9be5-cb569b373191 for instance with vm_state building and task_state spawning. [ 865.631741] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.632333] env[61911]: DEBUG nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 865.635458] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.197s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.635703] env[61911]: DEBUG nova.objects.instance [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lazy-loading 'resources' on Instance uuid 6e100d92-eb95-4b9b-b236-34ff94051811 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.739691] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 865.740394] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-288fe61b-5191-42bc-8d90-32268b11a8ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.751143] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 865.751143] env[61911]: value = "task-1251159" [ 865.751143] env[61911]: _type = "Task" [ 865.751143] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.762904] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251159, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.842364] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.845019] env[61911]: DEBUG nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Instance network_info: |[{"id": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "address": "fa:16:3e:3d:35:a7", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54a15f3e-c3", "ovs_interfaceid": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 865.845019] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:35:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54a15f3e-c3d1-45b1-ac13-19e74bed26c6', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.851688] env[61911]: DEBUG oslo.service.loopingcall [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.852065] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.852431] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b39a450-e299-482f-97b9-dedd8e358c52 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.873739] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.873739] env[61911]: value = "task-1251160" [ 865.873739] env[61911]: _type = "Task" [ 865.873739] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.883239] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251160, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.966037] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529df7dc-1dbf-a9b2-f0b7-23dbfaf30669, 'name': SearchDatastore_Task, 'duration_secs': 0.012388} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.966037] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.966551] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.967082] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.970149] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.970149] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.970149] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cf476bf-240b-467a-8aab-d21bed1295cf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.979554] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.980086] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.981676] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a97bf77-608a-4aab-95af-c3878a44d912 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.988066] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3eacaff9-b554-4f95-a973-589079e72227 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.097s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.990621] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 865.990621] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb7e5f-f203-fe69-e8c7-01a06c90c136" [ 865.990621] env[61911]: _type = "Task" [ 865.990621] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.005207] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.005207] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.005207] env[61911]: DEBUG nova.network.neutron [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.005207] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb7e5f-f203-fe69-e8c7-01a06c90c136, 'name': SearchDatastore_Task, 'duration_secs': 0.009639} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.005666] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71c77570-aadc-40af-9e89-5b1772c4a256 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.013332] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 866.013332] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de2eb4-354c-48d5-cc9c-7b7f64eae988" [ 866.013332] env[61911]: _type = "Task" [ 866.013332] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.022552] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de2eb4-354c-48d5-cc9c-7b7f64eae988, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.045734] env[61911]: DEBUG nova.compute.manager [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Received event network-changed-54a15f3e-c3d1-45b1-ac13-19e74bed26c6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 866.046233] env[61911]: DEBUG nova.compute.manager [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Refreshing instance network info cache due to event network-changed-54a15f3e-c3d1-45b1-ac13-19e74bed26c6. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 866.046233] env[61911]: DEBUG oslo_concurrency.lockutils [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] Acquiring lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.046373] env[61911]: DEBUG oslo_concurrency.lockutils [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] Acquired lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.046476] env[61911]: DEBUG nova.network.neutron [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Refreshing network info cache for port 54a15f3e-c3d1-45b1-ac13-19e74bed26c6 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.144967] env[61911]: DEBUG nova.compute.utils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.149032] env[61911]: DEBUG nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 866.149032] env[61911]: DEBUG nova.network.neutron [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.203029] env[61911]: DEBUG nova.policy [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de88f0838e5f471b9c67f8d85d21a87f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3fa4a98db1e4f39831a88402ec1ef6a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.263964] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251159, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.394049] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.394948] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.396454] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251160, 'name': CreateVM_Task, 'duration_secs': 0.327966} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.397504] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.399514] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.399869] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.400295] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.400698] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb9dc99d-537f-4b11-9295-b3e609319a5e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.411047] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 866.411047] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52428d29-bd65-e00f-4183-cfb0abfc2e04" [ 866.411047] env[61911]: _type = "Task" [ 866.411047] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.428704] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52428d29-bd65-e00f-4183-cfb0abfc2e04, 'name': SearchDatastore_Task, 'duration_secs': 0.011742} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.428704] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.428704] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.428704] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.494029] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b572fe83-92e8-4e90-9865-e137b3c84301 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.496387] env[61911]: DEBUG nova.network.neutron [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Successfully created port: 6e7b50c2-7b5e-4b94-8267-825879051466 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.509837] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160393f9-7531-41aa-908e-45d80158f8f4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.549354] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c038c23-317f-444f-bb70-9551e206a16e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.557451] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de2eb4-354c-48d5-cc9c-7b7f64eae988, 'name': SearchDatastore_Task, 'duration_secs': 0.01059} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.558413] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.558681] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.559063] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.559269] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.559496] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28641e7b-d795-4c9b-880f-f96e6e4e9f7e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.564042] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c2ab65f-49ce-44ce-89d0-441750f1d3ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.568175] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d494f8-eda5-4d14-adfc-0cce1b3c3833 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.573798] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 866.573798] env[61911]: value = "task-1251161" [ 866.573798] env[61911]: _type = "Task" [ 866.573798] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.589019] env[61911]: DEBUG nova.network.neutron [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.590711] env[61911]: DEBUG nova.compute.provider_tree [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.593460] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.593642] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.594574] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1255f99a-677c-44fb-a0ff-123f0d613065 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.599659] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251161, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.603423] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 866.603423] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523550c6-601f-6487-c6a5-e0b9a710b1aa" [ 866.603423] env[61911]: _type = "Task" [ 866.603423] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.612055] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523550c6-601f-6487-c6a5-e0b9a710b1aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.646974] env[61911]: DEBUG nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 866.762789] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251159, 'name': CreateSnapshot_Task, 'duration_secs': 0.696034} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.765462] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 866.766488] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757c203d-af20-4ade-9aa7-5f5f6e423e53 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.903436] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.903632] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 866.903756] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Rebuilding the list of instances to heal {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 867.086512] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251161, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.095613] env[61911]: DEBUG nova.scheduler.client.report [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 867.113902] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523550c6-601f-6487-c6a5-e0b9a710b1aa, 'name': SearchDatastore_Task, 'duration_secs': 0.009417} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.114681] env[61911]: DEBUG nova.network.neutron [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Updating instance_info_cache with network_info: [{"id": "0e16550a-783a-473b-9be5-cb569b373191", "address": "fa:16:3e:a7:ad:f8", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e16550a-78", "ovs_interfaceid": "0e16550a-783a-473b-9be5-cb569b373191", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.119028] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11d16549-0cde-40f8-8880-aa402d2600a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.122622] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 867.122622] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52783dca-c1a5-c13d-3349-abfedc1d8281" [ 867.122622] env[61911]: _type = "Task" [ 867.122622] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.138715] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52783dca-c1a5-c13d-3349-abfedc1d8281, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.253593] env[61911]: DEBUG nova.network.neutron [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Updated VIF entry in instance network info cache for port 54a15f3e-c3d1-45b1-ac13-19e74bed26c6. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 867.253954] env[61911]: DEBUG nova.network.neutron [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Updating instance_info_cache with network_info: [{"id": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "address": "fa:16:3e:3d:35:a7", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54a15f3e-c3", "ovs_interfaceid": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.293956] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 867.294279] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-91d8729d-31a7-4e1c-9fcc-1aec211d0a64 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.306222] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 867.306222] env[61911]: value = "task-1251162" [ 867.306222] env[61911]: _type = "Task" [ 867.306222] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.318135] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251162, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.409900] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 867.410114] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 867.410202] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Skipping network cache update for instance because it is Building. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 867.459551] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.459740] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.459945] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 867.460166] env[61911]: DEBUG nova.objects.instance [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lazy-loading 'info_cache' on Instance uuid 66472f43-537d-4eb3-8d49-d40627a8809d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.545747] env[61911]: DEBUG nova.compute.manager [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Received event network-changed-0e16550a-783a-473b-9be5-cb569b373191 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 867.545997] env[61911]: DEBUG nova.compute.manager [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Refreshing instance network info cache due to event network-changed-0e16550a-783a-473b-9be5-cb569b373191. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 867.546237] env[61911]: DEBUG oslo_concurrency.lockutils [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] Acquiring lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.586098] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251161, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554552} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.586375] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.586601] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.586927] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9b7d7d4-37d6-426b-bde4-96c52dba2ddb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.593988] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 867.593988] env[61911]: value = "task-1251163" [ 867.593988] env[61911]: _type = "Task" [ 867.593988] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.604172] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.968s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.605936] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251163, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.606456] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.878s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.606764] env[61911]: DEBUG nova.objects.instance [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lazy-loading 'resources' on Instance uuid 4b113b46-1da2-49ad-a0bb-c01b56c70f00 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.617679] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.618050] env[61911]: DEBUG nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Instance network_info: |[{"id": "0e16550a-783a-473b-9be5-cb569b373191", "address": "fa:16:3e:a7:ad:f8", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e16550a-78", "ovs_interfaceid": "0e16550a-783a-473b-9be5-cb569b373191", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 867.618768] env[61911]: DEBUG oslo_concurrency.lockutils [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] Acquired lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.618768] env[61911]: DEBUG nova.network.neutron [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Refreshing network info cache for port 0e16550a-783a-473b-9be5-cb569b373191 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.620423] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:ad:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '098df9b7-d759-47f7-b756-334848cb423b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e16550a-783a-473b-9be5-cb569b373191', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.629665] env[61911]: DEBUG oslo.service.loopingcall [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.630900] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.635272] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5810d6c-ec41-42ef-b329-ca5e1aeebcf5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.651147] env[61911]: INFO nova.scheduler.client.report [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Deleted allocations for instance 6e100d92-eb95-4b9b-b236-34ff94051811 [ 867.659230] env[61911]: DEBUG nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 867.670024] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52783dca-c1a5-c13d-3349-abfedc1d8281, 'name': SearchDatastore_Task, 'duration_secs': 0.012815} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.670024] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.670024] env[61911]: value = "task-1251164" [ 867.670024] env[61911]: _type = "Task" [ 867.670024] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.670024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.670024] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 9c2e75d8-b307-43bb-84a8-604cd5edbc39/9c2e75d8-b307-43bb-84a8-604cd5edbc39.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.670024] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36c7ac72-7df6-43b3-936e-8d3d2a71cde6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.677773] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251164, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.681458] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 867.681458] env[61911]: value = "task-1251165" [ 867.681458] env[61911]: _type = "Task" [ 867.681458] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.689110] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 867.689412] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.689581] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 867.689769] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.689982] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 867.690106] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 867.690473] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 867.690473] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 867.691082] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 867.691082] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 867.691082] env[61911]: DEBUG nova.virt.hardware [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 867.692498] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e93265-8994-41f3-92ed-7bf9708533c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.698390] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.704511] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0501b097-c7b9-4413-9e45-40a4383fceea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.757174] env[61911]: DEBUG oslo_concurrency.lockutils [req-be846fec-4104-47bf-b5b4-c81c399299f2 req-89f2fa46-f46c-4a03-aea5-e4928c00760d service nova] Releasing lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.821238] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251162, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.067132] env[61911]: DEBUG nova.network.neutron [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Successfully updated port: 6e7b50c2-7b5e-4b94-8267-825879051466 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.121257] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251163, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064968} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.123205] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.124825] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95df7df8-a7a4-4f00-9457-6e9cfaf50ed6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.179023] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.180503] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e13245bc-d989-4773-8d40-5c453f431140 tempest-ServerGroupTestJSON-1607932181 tempest-ServerGroupTestJSON-1607932181-project-member] Lock "6e100d92-eb95-4b9b-b236-34ff94051811" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.305s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.185475] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74454e26-2ce6-40f8-bfb8-3a883eeca3b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.204869] env[61911]: DEBUG nova.compute.manager [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 868.205100] env[61911]: DEBUG nova.compute.manager [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing instance network info cache due to event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 868.205268] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.205420] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.205618] env[61911]: DEBUG nova.network.neutron [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.228118] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251164, 'name': CreateVM_Task, 'duration_secs': 0.420878} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.234977] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.235596] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 868.235596] env[61911]: value = "task-1251166" [ 868.235596] env[61911]: _type = "Task" [ 868.235596] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.236078] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251165, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.239756] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.239925] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.240504] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.242549] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0451c2f-6e8b-4297-b8a3-3fdb4fb9a5e1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.255431] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251166, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.260678] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 868.260678] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5212566f-91d8-e649-d584-c94f04009173" [ 868.260678] env[61911]: _type = "Task" [ 868.260678] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.276040] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5212566f-91d8-e649-d584-c94f04009173, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.322835] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251162, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.508424] env[61911]: DEBUG nova.network.neutron [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Updated VIF entry in instance network info cache for port 0e16550a-783a-473b-9be5-cb569b373191. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.508785] env[61911]: DEBUG nova.network.neutron [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Updating instance_info_cache with network_info: [{"id": "0e16550a-783a-473b-9be5-cb569b373191", "address": "fa:16:3e:a7:ad:f8", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e16550a-78", "ovs_interfaceid": "0e16550a-783a-473b-9be5-cb569b373191", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.557034] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355be6b4-f6ee-45fe-b7f7-b169dcd226fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.568395] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154f87ef-7a3e-40f8-b5aa-e86f5da08150 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.572730] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "refresh_cache-b6b16408-7633-4f23-8f57-1193b1fb0384" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.572730] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired lock "refresh_cache-b6b16408-7633-4f23-8f57-1193b1fb0384" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.572730] env[61911]: DEBUG nova.network.neutron [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.611285] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c4978d-ae5d-4a1e-aaee-f9dd25e89afb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.620071] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed869b35-10c9-423f-958e-4f2d9235735e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.635486] env[61911]: DEBUG nova.compute.provider_tree [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.723607] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.726256} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.723607] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 9c2e75d8-b307-43bb-84a8-604cd5edbc39/9c2e75d8-b307-43bb-84a8-604cd5edbc39.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.723607] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.723607] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bdc38877-5fd2-4aee-aac6-ec47d5559f5f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.729950] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 868.729950] env[61911]: value = "task-1251167" [ 868.729950] env[61911]: _type = "Task" [ 868.729950] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.740334] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251167, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.750638] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251166, 'name': ReconfigVM_Task, 'duration_secs': 0.36805} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.750903] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 9230cd85-960d-427c-b69c-d19ac2bcfe35/9230cd85-960d-427c-b69c-d19ac2bcfe35.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.751563] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-817d940d-56be-46e3-9866-067a09bba24a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.759330] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 868.759330] env[61911]: value = "task-1251168" [ 868.759330] env[61911]: _type = "Task" [ 868.759330] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.783376] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5212566f-91d8-e649-d584-c94f04009173, 'name': SearchDatastore_Task, 'duration_secs': 0.061719} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.786836] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.787289] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.787548] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.787705] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.787904] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.788222] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251168, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.791172] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94f1e886-f9e1-4b42-b3d1-e7f33243b870 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.805585] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.805791] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.806752] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f7240f8-8a51-4cc8-a8bc-400aa2b86f19 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.817716] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 868.817716] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525dd31c-7cf0-d13a-6122-13ef0c74f59d" [ 868.817716] env[61911]: _type = "Task" [ 868.817716] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.821625] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251162, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.830774] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525dd31c-7cf0-d13a-6122-13ef0c74f59d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.012048] env[61911]: DEBUG oslo_concurrency.lockutils [req-45748d35-6de9-4881-b1db-9a1af8cd7c56 req-584e89d2-c7f0-4090-a75b-1c395b381117 service nova] Releasing lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.139904] env[61911]: DEBUG nova.scheduler.client.report [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.171748] env[61911]: DEBUG nova.network.neutron [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.244121] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251167, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067631} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.244121] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.245192] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dffcb4b-827c-4e9c-87bb-c75f7e1c7de5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.276853] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 9c2e75d8-b307-43bb-84a8-604cd5edbc39/9c2e75d8-b307-43bb-84a8-604cd5edbc39.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.284360] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6deec6e9-9e4b-478b-aabe-9064d1dc4257 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.312179] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251168, 'name': Rename_Task, 'duration_secs': 0.171619} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.321017] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.321017] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 869.321017] env[61911]: value = "task-1251169" [ 869.321017] env[61911]: _type = "Task" [ 869.321017] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.321017] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c2dd038-090d-4116-8e1c-cf4103c9783a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.328186] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251162, 'name': CloneVM_Task} progress is 95%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.339250] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 869.339250] env[61911]: value = "task-1251170" [ 869.339250] env[61911]: _type = "Task" [ 869.339250] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.342763] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251169, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.350204] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525dd31c-7cf0-d13a-6122-13ef0c74f59d, 'name': SearchDatastore_Task, 'duration_secs': 0.017536} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.351594] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a133ce61-215e-4811-b344-2322906c27fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.357796] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251170, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.363597] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 869.363597] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525c8996-5c4e-6b79-72ad-374155c87910" [ 869.363597] env[61911]: _type = "Task" [ 869.363597] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.371460] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525c8996-5c4e-6b79-72ad-374155c87910, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.573119] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Updating instance_info_cache with network_info: [{"id": "90825141-c63e-4853-8f27-6553cc0f3ecc", "address": "fa:16:3e:94:fb:80", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90825141-c6", "ovs_interfaceid": "90825141-c63e-4853-8f27-6553cc0f3ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.643707] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.037s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.646396] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.878s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.648201] env[61911]: INFO nova.compute.claims [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.664450] env[61911]: DEBUG nova.network.neutron [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Updating instance_info_cache with network_info: [{"id": "6e7b50c2-7b5e-4b94-8267-825879051466", "address": "fa:16:3e:a0:0e:79", "network": {"id": "08a38fc2-219a-4060-849d-9cc2f226d818", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1686086958-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3fa4a98db1e4f39831a88402ec1ef6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e7b50c2-7b", "ovs_interfaceid": "6e7b50c2-7b5e-4b94-8267-825879051466", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.669155] env[61911]: INFO nova.scheduler.client.report [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleted allocations for instance 4b113b46-1da2-49ad-a0bb-c01b56c70f00 [ 869.718816] env[61911]: DEBUG nova.network.neutron [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updated VIF entry in instance network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.718870] env[61911]: DEBUG nova.network.neutron [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.739000] env[61911]: DEBUG nova.compute.manager [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Received event network-vif-plugged-6e7b50c2-7b5e-4b94-8267-825879051466 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 869.739226] env[61911]: DEBUG oslo_concurrency.lockutils [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] Acquiring lock "b6b16408-7633-4f23-8f57-1193b1fb0384-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.739599] env[61911]: DEBUG oslo_concurrency.lockutils [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.740149] env[61911]: DEBUG oslo_concurrency.lockutils [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.740240] env[61911]: DEBUG nova.compute.manager [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] No waiting events found dispatching network-vif-plugged-6e7b50c2-7b5e-4b94-8267-825879051466 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 869.740449] env[61911]: WARNING nova.compute.manager [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Received unexpected event network-vif-plugged-6e7b50c2-7b5e-4b94-8267-825879051466 for instance with vm_state building and task_state spawning. [ 869.741814] env[61911]: DEBUG nova.compute.manager [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Received event network-changed-6e7b50c2-7b5e-4b94-8267-825879051466 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 869.741814] env[61911]: DEBUG nova.compute.manager [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Refreshing instance network info cache due to event network-changed-6e7b50c2-7b5e-4b94-8267-825879051466. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 869.741929] env[61911]: DEBUG oslo_concurrency.lockutils [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] Acquiring lock "refresh_cache-b6b16408-7633-4f23-8f57-1193b1fb0384" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.831276] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251162, 'name': CloneVM_Task, 'duration_secs': 2.220306} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.832208] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Created linked-clone VM from snapshot [ 869.833613] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff568013-a762-4692-93e3-c9a3465f7f32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.840655] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251169, 'name': ReconfigVM_Task, 'duration_secs': 0.384204} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.841544] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 9c2e75d8-b307-43bb-84a8-604cd5edbc39/9c2e75d8-b307-43bb-84a8-604cd5edbc39.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.842777] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66509ca2-434f-42ef-891c-c879d23f4035 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.848283] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Uploading image 37eeba24-1178-4ae0-aa71-cb4902ea42c8 {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 869.861386] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251170, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.863249] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 869.863249] env[61911]: value = "task-1251171" [ 869.863249] env[61911]: _type = "Task" [ 869.863249] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.878749] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525c8996-5c4e-6b79-72ad-374155c87910, 'name': SearchDatastore_Task, 'duration_secs': 0.019486} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.884103] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 869.884103] env[61911]: value = "vm-269606" [ 869.884103] env[61911]: _type = "VirtualMachine" [ 869.884103] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 869.884553] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.885335] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/1058f8d2-0979-4423-802d-52d91d5a2ff7.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 869.885335] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251171, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.885874] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-83c659cb-a61e-4911-9d68-1ee2a8c85559 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.888867] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10a0464f-6d26-40a1-858c-3b71f1e94837 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.899288] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 869.899288] env[61911]: value = "task-1251172" [ 869.899288] env[61911]: _type = "Task" [ 869.899288] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.900452] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lease: (returnval){ [ 869.900452] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526335e9-9f52-6e1d-6bb1-d36a2d1d5d68" [ 869.900452] env[61911]: _type = "HttpNfcLease" [ 869.900452] env[61911]: } obtained for exporting VM: (result){ [ 869.900452] env[61911]: value = "vm-269606" [ 869.900452] env[61911]: _type = "VirtualMachine" [ 869.900452] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 869.900678] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the lease: (returnval){ [ 869.900678] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526335e9-9f52-6e1d-6bb1-d36a2d1d5d68" [ 869.900678] env[61911]: _type = "HttpNfcLease" [ 869.900678] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 869.913401] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 869.913401] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526335e9-9f52-6e1d-6bb1-d36a2d1d5d68" [ 869.913401] env[61911]: _type = "HttpNfcLease" [ 869.913401] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 869.916675] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251172, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.075752] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-66472f43-537d-4eb3-8d49-d40627a8809d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.075964] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 870.076198] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.076360] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.076503] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.076649] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.076786] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.077047] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.077144] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 870.077285] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.167783] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Releasing lock "refresh_cache-b6b16408-7633-4f23-8f57-1193b1fb0384" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.168169] env[61911]: DEBUG nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Instance network_info: |[{"id": "6e7b50c2-7b5e-4b94-8267-825879051466", "address": "fa:16:3e:a0:0e:79", "network": {"id": "08a38fc2-219a-4060-849d-9cc2f226d818", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1686086958-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3fa4a98db1e4f39831a88402ec1ef6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e7b50c2-7b", "ovs_interfaceid": "6e7b50c2-7b5e-4b94-8267-825879051466", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 870.168715] env[61911]: DEBUG oslo_concurrency.lockutils [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] Acquired lock "refresh_cache-b6b16408-7633-4f23-8f57-1193b1fb0384" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.169659] env[61911]: DEBUG nova.network.neutron [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Refreshing network info cache for port 6e7b50c2-7b5e-4b94-8267-825879051466 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.170787] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:0e:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec763be6-4041-4651-8fd7-3820cf0ab86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e7b50c2-7b5e-4b94-8267-825879051466', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.179141] env[61911]: DEBUG oslo.service.loopingcall [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.182523] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.183054] env[61911]: DEBUG oslo_concurrency.lockutils [None req-357cfa2a-5a4f-45be-9dd9-6f488409745e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "4b113b46-1da2-49ad-a0bb-c01b56c70f00" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.156s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.184090] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ca6c922-80e5-4b71-94c9-47c5c39bb03d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.208887] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.208887] env[61911]: value = "task-1251174" [ 870.208887] env[61911]: _type = "Task" [ 870.208887] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.216250] env[61911]: DEBUG nova.compute.manager [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 870.216250] env[61911]: DEBUG nova.compute.manager [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing instance network info cache due to event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 870.216359] env[61911]: DEBUG oslo_concurrency.lockutils [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.216823] env[61911]: DEBUG oslo_concurrency.lockutils [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.216823] env[61911]: DEBUG nova.network.neutron [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.222243] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251174, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.225793] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.225793] env[61911]: DEBUG nova.compute.manager [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 870.225793] env[61911]: DEBUG nova.compute.manager [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing instance network info cache due to event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 870.225793] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.322951] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "interface-e9833b95-4162-42ba-87a4-d4cc790ac8e5-c34b4a10-4268-487c-bb19-1ef7beda1f06" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.323324] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e9833b95-4162-42ba-87a4-d4cc790ac8e5-c34b4a10-4268-487c-bb19-1ef7beda1f06" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.323689] env[61911]: DEBUG nova.objects.instance [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lazy-loading 'flavor' on Instance uuid e9833b95-4162-42ba-87a4-d4cc790ac8e5 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.365530] env[61911]: DEBUG oslo_vmware.api [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251170, 'name': PowerOnVM_Task, 'duration_secs': 0.528047} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.366778] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.366778] env[61911]: DEBUG nova.compute.manager [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 870.367096] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359fb661-172c-4cc6-befd-813411735494 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.383810] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251171, 'name': Rename_Task, 'duration_secs': 0.169858} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.385892] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.390483] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3484d423-35f9-4b06-9f8f-090ade074e02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.400953] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 870.400953] env[61911]: value = "task-1251175" [ 870.400953] env[61911]: _type = "Task" [ 870.400953] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.421734] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251172, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.422019] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 870.422019] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526335e9-9f52-6e1d-6bb1-d36a2d1d5d68" [ 870.422019] env[61911]: _type = "HttpNfcLease" [ 870.422019] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 870.425856] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 870.425856] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526335e9-9f52-6e1d-6bb1-d36a2d1d5d68" [ 870.425856] env[61911]: _type = "HttpNfcLease" [ 870.425856] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 870.427050] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251175, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.427050] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed87032-344c-4051-9b9d-2f991bdeac12 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.437446] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526960f3-2a00-c74a-ed15-a006ff961ee6/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 870.437446] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526960f3-2a00-c74a-ed15-a006ff961ee6/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 870.580453] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.586183] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-09287d30-84c4-4f57-9687-7db71aa4eb2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.719948] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251174, 'name': CreateVM_Task, 'duration_secs': 0.449652} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.720197] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.721094] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.721248] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.721578] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.724092] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e755a7e-3b2f-47d9-bb0d-212047cb4703 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.731547] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 870.731547] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e791c0-227d-df28-e0a9-76af2e85b085" [ 870.731547] env[61911]: _type = "Task" [ 870.731547] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.743553] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e791c0-227d-df28-e0a9-76af2e85b085, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.911405] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.924406] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251172, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.701193} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.930333] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/1058f8d2-0979-4423-802d-52d91d5a2ff7.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.930680] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.931029] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251175, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.932705] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57c7d1f7-ba43-4397-99b3-2dc70b623d1e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.941466] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 870.941466] env[61911]: value = "task-1251176" [ 870.941466] env[61911]: _type = "Task" [ 870.941466] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.948992] env[61911]: DEBUG nova.objects.instance [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lazy-loading 'pci_requests' on Instance uuid e9833b95-4162-42ba-87a4-d4cc790ac8e5 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.960019] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251176, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.000229] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cd7c48-ef60-4a7b-8336-5916470254e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.010496] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fb533c-b0d9-41cf-b32e-e61d861a39cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.053099] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f54db82-0fa0-44ee-9f14-43f857c111c9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.065090] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30124b79-9836-4e73-b063-ca2fc30646ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.080867] env[61911]: DEBUG nova.compute.provider_tree [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.174384] env[61911]: DEBUG nova.network.neutron [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Updated VIF entry in instance network info cache for port 6e7b50c2-7b5e-4b94-8267-825879051466. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.175072] env[61911]: DEBUG nova.network.neutron [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Updating instance_info_cache with network_info: [{"id": "6e7b50c2-7b5e-4b94-8267-825879051466", "address": "fa:16:3e:a0:0e:79", "network": {"id": "08a38fc2-219a-4060-849d-9cc2f226d818", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1686086958-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3fa4a98db1e4f39831a88402ec1ef6a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec763be6-4041-4651-8fd7-3820cf0ab86d", "external-id": "nsx-vlan-transportzone-943", "segmentation_id": 943, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e7b50c2-7b", "ovs_interfaceid": "6e7b50c2-7b5e-4b94-8267-825879051466", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.239482] env[61911]: DEBUG nova.network.neutron [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updated VIF entry in instance network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.239846] env[61911]: DEBUG nova.network.neutron [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.245744] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e791c0-227d-df28-e0a9-76af2e85b085, 'name': SearchDatastore_Task, 'duration_secs': 0.012701} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.246426] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.246628] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.246971] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.247939] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.247939] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.247939] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bcfe7a95-e326-4f7a-8dc9-fc323af2c380 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.259610] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.259892] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.260755] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc15c143-4300-4292-ae0e-617b06819d06 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.267160] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 871.267160] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5238f975-c052-8b55-f699-a2d73187f1fa" [ 871.267160] env[61911]: _type = "Task" [ 871.267160] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.275929] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5238f975-c052-8b55-f699-a2d73187f1fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.422076] env[61911]: DEBUG oslo_vmware.api [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251175, 'name': PowerOnVM_Task, 'duration_secs': 0.553683} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.422415] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.422635] env[61911]: INFO nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Took 8.75 seconds to spawn the instance on the hypervisor. [ 871.422871] env[61911]: DEBUG nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 871.423677] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1264b169-26f3-4288-a2e2-cbed1129d21b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.452766] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086678} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.453503] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.455058] env[61911]: DEBUG nova.objects.base [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 871.455058] env[61911]: DEBUG nova.network.neutron [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 871.458869] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cbcac1-817d-442d-a20b-7495ef076f2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.488885] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/1058f8d2-0979-4423-802d-52d91d5a2ff7.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.490181] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a96c2a0-a887-4272-9424-24d4060d8d6b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.515049] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 871.515049] env[61911]: value = "task-1251177" [ 871.515049] env[61911]: _type = "Task" [ 871.515049] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.524826] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.575021] env[61911]: DEBUG nova.policy [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ee96b52414148b7bff1d067595848b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f40f512384d4fedb75bfc53f3195243', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.583591] env[61911]: DEBUG nova.scheduler.client.report [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.678261] env[61911]: DEBUG oslo_concurrency.lockutils [req-16dffacc-ad51-441e-b61d-79da52b32918 req-32facf06-202f-420d-9c8f-9359282c6a09 service nova] Releasing lock "refresh_cache-b6b16408-7633-4f23-8f57-1193b1fb0384" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.747458] env[61911]: DEBUG oslo_concurrency.lockutils [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.747778] env[61911]: DEBUG nova.compute.manager [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 871.747969] env[61911]: DEBUG nova.compute.manager [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing instance network info cache due to event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 871.748219] env[61911]: DEBUG oslo_concurrency.lockutils [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.748369] env[61911]: DEBUG oslo_concurrency.lockutils [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.748552] env[61911]: DEBUG nova.network.neutron [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.751836] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.752391] env[61911]: DEBUG nova.network.neutron [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.780181] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5238f975-c052-8b55-f699-a2d73187f1fa, 'name': SearchDatastore_Task, 'duration_secs': 0.023844} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.781098] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f41c2d8-dbec-4fb5-bd35-5fac77534395 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.788661] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 871.788661] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b282ec-8434-dedb-137c-1b66d889c04e" [ 871.788661] env[61911]: _type = "Task" [ 871.788661] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.797859] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b282ec-8434-dedb-137c-1b66d889c04e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.942955] env[61911]: INFO nova.compute.manager [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Took 28.42 seconds to build instance. [ 872.027587] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251177, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.090033] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.090769] env[61911]: DEBUG nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 872.096061] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.665s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.096401] env[61911]: DEBUG nova.objects.instance [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lazy-loading 'resources' on Instance uuid 63d54f69-3d8f-45e9-b317-9d484af954b1 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.149312] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "9230cd85-960d-427c-b69c-d19ac2bcfe35" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.149312] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.149312] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "9230cd85-960d-427c-b69c-d19ac2bcfe35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.149312] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.149312] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.152463] env[61911]: INFO nova.compute.manager [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Terminating instance [ 872.301554] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b282ec-8434-dedb-137c-1b66d889c04e, 'name': SearchDatastore_Task, 'duration_secs': 0.02193} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.304339] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.304607] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] b6b16408-7633-4f23-8f57-1193b1fb0384/b6b16408-7633-4f23-8f57-1193b1fb0384.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.309024] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e7fd913-0bcf-4c35-8be6-7c5a47104e91 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.316347] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 872.316347] env[61911]: value = "task-1251178" [ 872.316347] env[61911]: _type = "Task" [ 872.316347] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.330608] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251178, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.445189] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9671516c-6537-4ac2-b477-3afa9c3bf20b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.926s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.526735] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251177, 'name': ReconfigVM_Task, 'duration_secs': 0.691286} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.527901] env[61911]: DEBUG nova.network.neutron [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updated VIF entry in instance network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.528270] env[61911]: DEBUG nova.network.neutron [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.532263] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/1058f8d2-0979-4423-802d-52d91d5a2ff7.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.532902] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7069993b-0747-485b-9863-b87853a5523f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.543438] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 872.543438] env[61911]: value = "task-1251179" [ 872.543438] env[61911]: _type = "Task" [ 872.543438] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.558518] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251179, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.586056] env[61911]: DEBUG nova.network.neutron [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updated VIF entry in instance network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.586056] env[61911]: DEBUG nova.network.neutron [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.600402] env[61911]: DEBUG nova.compute.utils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.604678] env[61911]: DEBUG nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 872.604919] env[61911]: DEBUG nova.network.neutron [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.657639] env[61911]: DEBUG nova.compute.manager [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 872.657639] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.658034] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c782d0c-5084-4acd-9991-73e68f37e2fd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.675249] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.677120] env[61911]: DEBUG nova.policy [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62586445653844078feeeb731bd16f62', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e6ca33655884f16b4c0a6fcdb31ec12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.678960] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ea8f523-f3c9-484c-9bf3-81ade31cd40e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.692854] env[61911]: DEBUG oslo_vmware.api [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 872.692854] env[61911]: value = "task-1251180" [ 872.692854] env[61911]: _type = "Task" [ 872.692854] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.710314] env[61911]: DEBUG oslo_vmware.api [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251180, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.828504] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251178, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.976250] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cf44d2-ec81-4136-ab7b-a7eb07061c6b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.985159] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b4ccd0-f018-4ab7-832f-8ed702df7199 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.018853] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08adf371-be62-4baf-9597-cf8ae20fbf20 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.027375] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d62cd23-5e84-49be-8493-45320ec43a73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.033148] env[61911]: DEBUG oslo_concurrency.lockutils [req-e6da0ce2-b764-4011-ba4f-5652435b444a req-d24f0258-877d-4bd8-9d00-298bc59880b7 service nova] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.043893] env[61911]: DEBUG nova.compute.provider_tree [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.054097] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251179, 'name': Rename_Task, 'duration_secs': 0.239475} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.057167] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.058315] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04a51bad-6b3f-4246-ac83-6ff3bcc33709 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.060712] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.060961] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.069488] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 873.069488] env[61911]: value = "task-1251181" [ 873.069488] env[61911]: _type = "Task" [ 873.069488] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.079579] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251181, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.089032] env[61911]: DEBUG oslo_concurrency.lockutils [req-251b7a5a-036a-4054-a085-e9cb804d1ca8 req-52739c81-aacb-48bd-97f1-d3834044535d service nova] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.108082] env[61911]: DEBUG nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 873.203087] env[61911]: DEBUG oslo_vmware.api [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251180, 'name': PowerOffVM_Task, 'duration_secs': 0.278836} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.203382] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.203553] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.203811] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf4c7a0d-a65b-46cb-af1f-508e2cd37887 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.285805] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.287054] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.287352] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleting the datastore file [datastore2] 9230cd85-960d-427c-b69c-d19ac2bcfe35 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.287652] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b4f3c6f-b2b2-4ae9-b7c6-5df5089a2d85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.299161] env[61911]: DEBUG oslo_vmware.api [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 873.299161] env[61911]: value = "task-1251183" [ 873.299161] env[61911]: _type = "Task" [ 873.299161] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.312768] env[61911]: DEBUG oslo_vmware.api [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.332475] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251178, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.379961] env[61911]: DEBUG nova.network.neutron [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Successfully created port: 5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.549352] env[61911]: DEBUG nova.scheduler.client.report [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 873.566256] env[61911]: DEBUG nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 873.586734] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251181, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.622861] env[61911]: DEBUG nova.network.neutron [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Successfully updated port: c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.811251] env[61911]: DEBUG oslo_vmware.api [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284248} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.811550] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.811672] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.811881] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.812234] env[61911]: INFO nova.compute.manager [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Took 1.16 seconds to destroy the instance on the hypervisor. [ 873.812593] env[61911]: DEBUG oslo.service.loopingcall [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.812848] env[61911]: DEBUG nova.compute.manager [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 873.812930] env[61911]: DEBUG nova.network.neutron [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.829138] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251178, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.018944] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] Acquiring lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.019251] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] Acquired lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.019591] env[61911]: DEBUG nova.network.neutron [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.045845] env[61911]: DEBUG nova.compute.manager [req-61173f08-d51a-446f-9012-2f2ac9383900 req-086642c5-e471-4ab2-b5ad-ae701c9db3a6 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-vif-plugged-c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 874.046152] env[61911]: DEBUG oslo_concurrency.lockutils [req-61173f08-d51a-446f-9012-2f2ac9383900 req-086642c5-e471-4ab2-b5ad-ae701c9db3a6 service nova] Acquiring lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.046344] env[61911]: DEBUG oslo_concurrency.lockutils [req-61173f08-d51a-446f-9012-2f2ac9383900 req-086642c5-e471-4ab2-b5ad-ae701c9db3a6 service nova] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.046535] env[61911]: DEBUG oslo_concurrency.lockutils [req-61173f08-d51a-446f-9012-2f2ac9383900 req-086642c5-e471-4ab2-b5ad-ae701c9db3a6 service nova] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.046697] env[61911]: DEBUG nova.compute.manager [req-61173f08-d51a-446f-9012-2f2ac9383900 req-086642c5-e471-4ab2-b5ad-ae701c9db3a6 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] No waiting events found dispatching network-vif-plugged-c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 874.046888] env[61911]: WARNING nova.compute.manager [req-61173f08-d51a-446f-9012-2f2ac9383900 req-086642c5-e471-4ab2-b5ad-ae701c9db3a6 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received unexpected event network-vif-plugged-c34b4a10-4268-487c-bb19-1ef7beda1f06 for instance with vm_state active and task_state None. [ 874.054732] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.057654] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.528s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.059192] env[61911]: INFO nova.compute.claims [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.084497] env[61911]: DEBUG oslo_vmware.api [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251181, 'name': PowerOnVM_Task, 'duration_secs': 0.716238} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.084886] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.085215] env[61911]: INFO nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Took 8.92 seconds to spawn the instance on the hypervisor. [ 874.085386] env[61911]: DEBUG nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 874.086318] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea60ac9c-842c-4a02-b13e-9c39cfaf608f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.090161] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.098993] env[61911]: INFO nova.scheduler.client.report [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted allocations for instance 63d54f69-3d8f-45e9-b317-9d484af954b1 [ 874.121710] env[61911]: DEBUG nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 874.129434] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.129703] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.129995] env[61911]: DEBUG nova.network.neutron [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.161412] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 874.162318] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.162318] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 874.162471] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.162697] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 874.163147] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 874.163246] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 874.163481] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 874.163758] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 874.164007] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 874.164326] env[61911]: DEBUG nova.virt.hardware [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 874.166941] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0afab8eb-527d-4dc3-8c10-64c9273fd747 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.178242] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fd6ba1-4b12-4de2-8fb2-c7f56fe8e5f5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.331079] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251178, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.588375} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.331385] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] b6b16408-7633-4f23-8f57-1193b1fb0384/b6b16408-7633-4f23-8f57-1193b1fb0384.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.331613] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.331872] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-983a90fa-be81-443d-9dcb-bcbe064ca9b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.340439] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 874.340439] env[61911]: value = "task-1251184" [ 874.340439] env[61911]: _type = "Task" [ 874.340439] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.351123] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.610015] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1332408d-1291-43d3-b85c-e375e2a428e0 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "63d54f69-3d8f-45e9-b317-9d484af954b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.636s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.611257] env[61911]: INFO nova.compute.manager [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Took 28.53 seconds to build instance. [ 874.632657] env[61911]: DEBUG nova.network.neutron [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.666973] env[61911]: WARNING nova.network.neutron [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] 87858d49-056a-4528-87e1-d07b16c7f333 already exists in list: networks containing: ['87858d49-056a-4528-87e1-d07b16c7f333']. ignoring it [ 874.851954] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111445} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.852193] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.852989] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24b7fae-20a2-47b9-93cc-eb9db7c255e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.878270] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] b6b16408-7633-4f23-8f57-1193b1fb0384/b6b16408-7633-4f23-8f57-1193b1fb0384.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.880896] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34b8a1f4-4347-4ee6-9338-3440a3a1fe84 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.901946] env[61911]: DEBUG nova.network.neutron [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Updating instance_info_cache with network_info: [{"id": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "address": "fa:16:3e:3d:35:a7", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54a15f3e-c3", "ovs_interfaceid": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.903932] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 874.903932] env[61911]: value = "task-1251185" [ 874.903932] env[61911]: _type = "Task" [ 874.903932] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.913567] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.055711] env[61911]: DEBUG nova.network.neutron [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "address": "fa:16:3e:b6:3a:b4", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc34b4a10-42", "ovs_interfaceid": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.114354] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23759845-8bb5-47e2-a4bb-ed0708455198 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.051s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.138287] env[61911]: INFO nova.compute.manager [-] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Took 1.32 seconds to deallocate network for instance. [ 875.348291] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4cdf1e-d016-4b83-ad18-ba9a0a4b58bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.356711] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e441dd35-0036-44c3-8389-98419b686717 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.388541] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a425041-394f-4e45-95d8-2ab56fd97119 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.397713] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bac7d2-012a-489a-b071-18c9fc5acfec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.412638] env[61911]: DEBUG oslo_concurrency.lockutils [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] Releasing lock "refresh_cache-9c2e75d8-b307-43bb-84a8-604cd5edbc39" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.412751] env[61911]: DEBUG nova.compute.manager [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Inject network info {{(pid=61911) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7599}} [ 875.413130] env[61911]: DEBUG nova.compute.manager [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] network_info to inject: |[{"id": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "address": "fa:16:3e:3d:35:a7", "network": {"id": "44f0ee49-584a-4d63-84bb-e15b0d3d87a0", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-388118607-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e38ba711d3e3431b9d258c91ee1ce6f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54a15f3e-c3", "ovs_interfaceid": "54a15f3e-c3d1-45b1-ac13-19e74bed26c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7600}} [ 875.417813] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Reconfiguring VM instance to set the machine id {{(pid=61911) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 875.418377] env[61911]: DEBUG nova.compute.provider_tree [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.419827] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73af4763-4159-4f2b-a92a-b6d00ba47b03 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.442429] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251185, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.444022] env[61911]: DEBUG oslo_vmware.api [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] Waiting for the task: (returnval){ [ 875.444022] env[61911]: value = "task-1251186" [ 875.444022] env[61911]: _type = "Task" [ 875.444022] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.452782] env[61911]: DEBUG oslo_vmware.api [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] Task: {'id': task-1251186, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.560283] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.561062] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.561187] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.562422] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d5f8c9-af64-489d-9f29-f7836aa03131 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.581987] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 875.582378] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.582618] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 875.582882] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.583125] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 875.583363] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 875.583647] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 875.583878] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 875.584148] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 875.584394] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 875.584650] env[61911]: DEBUG nova.virt.hardware [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 875.590988] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Reconfiguring VM to attach interface {{(pid=61911) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 875.591401] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2595fca9-dd71-4650-9437-62d8babbfbf9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.610220] env[61911]: DEBUG oslo_vmware.api [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 875.610220] env[61911]: value = "task-1251187" [ 875.610220] env[61911]: _type = "Task" [ 875.610220] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.620240] env[61911]: DEBUG oslo_vmware.api [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251187, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.645554] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.744388] env[61911]: DEBUG nova.network.neutron [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Successfully updated port: 5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.808230] env[61911]: DEBUG nova.compute.manager [req-f9f56de6-00de-4e66-a7e3-9b69707cbe22 req-b7ee7f29-1895-4fc5-8f8b-b8f1782a3489 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Received event network-vif-plugged-5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 875.808561] env[61911]: DEBUG oslo_concurrency.lockutils [req-f9f56de6-00de-4e66-a7e3-9b69707cbe22 req-b7ee7f29-1895-4fc5-8f8b-b8f1782a3489 service nova] Acquiring lock "5b7d09e1-ee14-4001-8d73-14e763402670-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.808885] env[61911]: DEBUG oslo_concurrency.lockutils [req-f9f56de6-00de-4e66-a7e3-9b69707cbe22 req-b7ee7f29-1895-4fc5-8f8b-b8f1782a3489 service nova] Lock "5b7d09e1-ee14-4001-8d73-14e763402670-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.809223] env[61911]: DEBUG oslo_concurrency.lockutils [req-f9f56de6-00de-4e66-a7e3-9b69707cbe22 req-b7ee7f29-1895-4fc5-8f8b-b8f1782a3489 service nova] Lock "5b7d09e1-ee14-4001-8d73-14e763402670-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.809563] env[61911]: DEBUG nova.compute.manager [req-f9f56de6-00de-4e66-a7e3-9b69707cbe22 req-b7ee7f29-1895-4fc5-8f8b-b8f1782a3489 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] No waiting events found dispatching network-vif-plugged-5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 875.809808] env[61911]: WARNING nova.compute.manager [req-f9f56de6-00de-4e66-a7e3-9b69707cbe22 req-b7ee7f29-1895-4fc5-8f8b-b8f1782a3489 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Received unexpected event network-vif-plugged-5c7df9be-477d-4aab-9d5b-e340866dc6bb for instance with vm_state building and task_state spawning. [ 875.929222] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251185, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.933620] env[61911]: DEBUG nova.scheduler.client.report [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 875.956135] env[61911]: DEBUG oslo_vmware.api [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] Task: {'id': task-1251186, 'name': ReconfigVM_Task, 'duration_secs': 0.245288} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.956438] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-85b272a8-922c-498f-93f5-21641460a2e8 tempest-ServersAdminTestJSON-882232182 tempest-ServersAdminTestJSON-882232182-project-admin] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Reconfigured VM instance to set the machine id {{(pid=61911) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 876.121339] env[61911]: DEBUG oslo_vmware.api [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.248573] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.248744] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.248901] env[61911]: DEBUG nova.network.neutron [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 876.399299] env[61911]: INFO nova.compute.manager [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Rescuing [ 876.399646] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.399883] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.400098] env[61911]: DEBUG nova.network.neutron [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 876.407050] env[61911]: DEBUG nova.compute.manager [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-changed-c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 876.407050] env[61911]: DEBUG nova.compute.manager [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing instance network info cache due to event network-changed-c34b4a10-4268-487c-bb19-1ef7beda1f06. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 876.407312] env[61911]: DEBUG oslo_concurrency.lockutils [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.407520] env[61911]: DEBUG oslo_concurrency.lockutils [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.407739] env[61911]: DEBUG nova.network.neutron [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing network info cache for port c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.432115] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251185, 'name': ReconfigVM_Task, 'duration_secs': 1.281438} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.432438] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Reconfigured VM instance instance-0000004a to attach disk [datastore2] b6b16408-7633-4f23-8f57-1193b1fb0384/b6b16408-7633-4f23-8f57-1193b1fb0384.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.433128] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-728a1720-dfb7-4896-ad39-64f81fa24970 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.439095] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.439611] env[61911]: DEBUG nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 876.443709] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.545s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.445320] env[61911]: INFO nova.compute.claims [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.450066] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 876.450066] env[61911]: value = "task-1251188" [ 876.450066] env[61911]: _type = "Task" [ 876.450066] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.460400] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251188, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.622859] env[61911]: DEBUG oslo_vmware.api [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251187, 'name': ReconfigVM_Task, 'duration_secs': 0.999299} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.623425] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.623703] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Reconfigured VM to attach interface {{(pid=61911) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 876.798307] env[61911]: DEBUG nova.network.neutron [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.954664] env[61911]: DEBUG nova.compute.utils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.956250] env[61911]: DEBUG nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 876.957075] env[61911]: DEBUG nova.network.neutron [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 876.972972] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251188, 'name': Rename_Task, 'duration_secs': 0.191693} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.976181] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.976181] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e23071b-144c-444d-b5c1-69e196c40cd1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.982077] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 876.982077] env[61911]: value = "task-1251189" [ 876.982077] env[61911]: _type = "Task" [ 876.982077] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.991799] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251189, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.028787] env[61911]: DEBUG nova.policy [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '012d3af62b60497a953855490ae89c42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '687cce9c9abb4c928295cbe60ec1d078', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.107967] env[61911]: DEBUG nova.network.neutron [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updating instance_info_cache with network_info: [{"id": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "address": "fa:16:3e:39:28:bf", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c7df9be-47", "ovs_interfaceid": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.128625] env[61911]: DEBUG oslo_concurrency.lockutils [None req-38a2446f-6bec-479f-a37b-aaebb7872d10 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e9833b95-4162-42ba-87a4-d4cc790ac8e5-c34b4a10-4268-487c-bb19-1ef7beda1f06" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.805s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.259543] env[61911]: DEBUG nova.network.neutron [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updated VIF entry in instance network info cache for port c34b4a10-4268-487c-bb19-1ef7beda1f06. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.260328] env[61911]: DEBUG nova.network.neutron [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "address": "fa:16:3e:b6:3a:b4", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc34b4a10-42", "ovs_interfaceid": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.332107] env[61911]: DEBUG nova.network.neutron [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Updating instance_info_cache with network_info: [{"id": "0e16550a-783a-473b-9be5-cb569b373191", "address": "fa:16:3e:a7:ad:f8", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e16550a-78", "ovs_interfaceid": "0e16550a-783a-473b-9be5-cb569b373191", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.467281] env[61911]: DEBUG nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 877.497719] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251189, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.501822] env[61911]: DEBUG nova.network.neutron [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Successfully created port: 82208eb4-5834-4277-9616-85003ea23d27 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.613657] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.613657] env[61911]: DEBUG nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Instance network_info: |[{"id": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "address": "fa:16:3e:39:28:bf", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c7df9be-47", "ovs_interfaceid": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 877.618201] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:28:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c7df9be-477d-4aab-9d5b-e340866dc6bb', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.626327] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Creating folder: Project (9e6ca33655884f16b4c0a6fcdb31ec12). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.626925] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af673f80-7d71-402a-ad31-a9927db505bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.643381] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Created folder: Project (9e6ca33655884f16b4c0a6fcdb31ec12) in parent group-v269521. [ 877.643759] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Creating folder: Instances. Parent ref: group-v269609. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 877.644016] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01e70c73-0fa5-4faa-bde1-e447ed6616a7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.659833] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Created folder: Instances in parent group-v269609. [ 877.660272] env[61911]: DEBUG oslo.service.loopingcall [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.660613] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.662097] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34c2099e-1240-47c1-bebd-a3edb0bc1fac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.687816] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.687816] env[61911]: value = "task-1251192" [ 877.687816] env[61911]: _type = "Task" [ 877.687816] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.699898] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251192, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.766830] env[61911]: DEBUG oslo_concurrency.lockutils [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.768010] env[61911]: DEBUG nova.compute.manager [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Received event network-vif-deleted-47a25084-ef5a-4059-92b8-d93d72091857 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 877.768010] env[61911]: DEBUG nova.compute.manager [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Received event network-changed-5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 877.768010] env[61911]: DEBUG nova.compute.manager [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Refreshing instance network info cache due to event network-changed-5c7df9be-477d-4aab-9d5b-e340866dc6bb. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 877.768010] env[61911]: DEBUG oslo_concurrency.lockutils [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] Acquiring lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.768010] env[61911]: DEBUG oslo_concurrency.lockutils [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] Acquired lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.768237] env[61911]: DEBUG nova.network.neutron [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Refreshing network info cache for port 5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.831954] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87938cb5-4e98-4995-a3fc-e9cdd10acaa6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.836810] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.843982] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fb4114-076d-4659-9f86-18bcc13a3ce8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.882229] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3599996-38df-4ccd-9f47-44538f0884e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.888115] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "50a522ab-651a-4bb3-93d3-8866169e95e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.888351] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.899754] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865af734-afe2-4165-a6b9-d25e0418cb23 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.917528] env[61911]: DEBUG nova.compute.provider_tree [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.001032] env[61911]: DEBUG oslo_vmware.api [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251189, 'name': PowerOnVM_Task, 'duration_secs': 0.669033} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.001032] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.001032] env[61911]: INFO nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Took 10.34 seconds to spawn the instance on the hypervisor. [ 878.001032] env[61911]: DEBUG nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 878.001032] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf81898-ab99-41f5-90e9-4927a508df5f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.199895] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251192, 'name': CreateVM_Task, 'duration_secs': 0.473816} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.200185] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.201050] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.201199] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.201572] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.201877] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99d06ac8-0c4e-4051-ae5a-2636183c9a37 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.207344] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 878.207344] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f466c5-6890-652c-8e76-f35a8c21a521" [ 878.207344] env[61911]: _type = "Task" [ 878.207344] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.225122] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f466c5-6890-652c-8e76-f35a8c21a521, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.390951] env[61911]: DEBUG nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 878.421461] env[61911]: DEBUG nova.scheduler.client.report [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 878.479430] env[61911]: DEBUG nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 878.511404] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 878.511841] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.512036] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 878.512237] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.512384] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 878.512530] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 878.512743] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 878.512901] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 878.513081] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 878.513248] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 878.513421] env[61911]: DEBUG nova.virt.hardware [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 878.518210] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833477d0-ced9-4d04-9ece-a0b0f5ce26cf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.523218] env[61911]: INFO nova.compute.manager [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Took 24.57 seconds to build instance. [ 878.529808] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946d9f13-99fd-4e53-a761-23825599bdde {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.553957] env[61911]: DEBUG nova.network.neutron [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updated VIF entry in instance network info cache for port 5c7df9be-477d-4aab-9d5b-e340866dc6bb. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.554373] env[61911]: DEBUG nova.network.neutron [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updating instance_info_cache with network_info: [{"id": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "address": "fa:16:3e:39:28:bf", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c7df9be-47", "ovs_interfaceid": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.720015] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f466c5-6890-652c-8e76-f35a8c21a521, 'name': SearchDatastore_Task, 'duration_secs': 0.01428} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.720357] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.720604] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.720879] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.720991] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.721195] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.721529] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b38f0558-2f9b-48d2-9397-21edb7e5a550 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.731456] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.731653] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.732426] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebb6ae37-6936-4d62-8484-98a684732449 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.738499] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 878.738499] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a031e5-c4be-0307-a546-9005992299ec" [ 878.738499] env[61911]: _type = "Task" [ 878.738499] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.747717] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a031e5-c4be-0307-a546-9005992299ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.924212] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.931369] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.931369] env[61911]: DEBUG nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 878.933372] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.352s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.934101] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.934101] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 878.934355] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.023s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.934576] env[61911]: DEBUG nova.objects.instance [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 878.942956] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f9a9ea-7470-4c14-9998-10f2eac7c6fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.959688] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b43a83-5c8c-4acc-96f4-d335ee700f9e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.968809] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "interface-e9833b95-4162-42ba-87a4-d4cc790ac8e5-c34b4a10-4268-487c-bb19-1ef7beda1f06" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.968809] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e9833b95-4162-42ba-87a4-d4cc790ac8e5-c34b4a10-4268-487c-bb19-1ef7beda1f06" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.989054] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf014ff-3583-4610-9e8e-0b8e84971c84 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.994920] env[61911]: DEBUG nova.compute.manager [req-8fe8f497-e2ef-4f85-939c-0437107b0ea7 req-e5c877de-dfe1-4a20-b34e-9392870b6655 service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Received event network-vif-plugged-82208eb4-5834-4277-9616-85003ea23d27 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 878.995356] env[61911]: DEBUG oslo_concurrency.lockutils [req-8fe8f497-e2ef-4f85-939c-0437107b0ea7 req-e5c877de-dfe1-4a20-b34e-9392870b6655 service nova] Acquiring lock "da6f7457-e603-415c-9578-024a7563d34f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.995735] env[61911]: DEBUG oslo_concurrency.lockutils [req-8fe8f497-e2ef-4f85-939c-0437107b0ea7 req-e5c877de-dfe1-4a20-b34e-9392870b6655 service nova] Lock "da6f7457-e603-415c-9578-024a7563d34f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.996038] env[61911]: DEBUG oslo_concurrency.lockutils [req-8fe8f497-e2ef-4f85-939c-0437107b0ea7 req-e5c877de-dfe1-4a20-b34e-9392870b6655 service nova] Lock "da6f7457-e603-415c-9578-024a7563d34f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.996351] env[61911]: DEBUG nova.compute.manager [req-8fe8f497-e2ef-4f85-939c-0437107b0ea7 req-e5c877de-dfe1-4a20-b34e-9392870b6655 service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] No waiting events found dispatching network-vif-plugged-82208eb4-5834-4277-9616-85003ea23d27 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 878.996616] env[61911]: WARNING nova.compute.manager [req-8fe8f497-e2ef-4f85-939c-0437107b0ea7 req-e5c877de-dfe1-4a20-b34e-9392870b6655 service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Received unexpected event network-vif-plugged-82208eb4-5834-4277-9616-85003ea23d27 for instance with vm_state building and task_state spawning. [ 879.004231] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aeff2ae-e739-4d31-97a6-2d2bfb99589e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.041765] env[61911]: DEBUG oslo_concurrency.lockutils [None req-93cd0390-91dd-4137-81a7-a5e44a9a16f5 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.103s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.042406] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179833MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 879.042611] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.057766] env[61911]: DEBUG oslo_concurrency.lockutils [req-af0652fb-b534-4928-97e2-c3939fb7659a req-055e0350-b6eb-41a4-8ab0-d2e04bd89457 service nova] Releasing lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.108478] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526960f3-2a00-c74a-ed15-a006ff961ee6/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 879.109413] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066107dc-ca5e-48a7-a3c1-0fc209fac90e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.116879] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526960f3-2a00-c74a-ed15-a006ff961ee6/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 879.117083] env[61911]: ERROR oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526960f3-2a00-c74a-ed15-a006ff961ee6/disk-0.vmdk due to incomplete transfer. [ 879.117355] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-089cfb36-9dd1-4c29-b70d-ecb151de9009 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.125561] env[61911]: DEBUG oslo_vmware.rw_handles [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526960f3-2a00-c74a-ed15-a006ff961ee6/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 879.125789] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Uploaded image 37eeba24-1178-4ae0-aa71-cb4902ea42c8 to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 879.128464] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 879.128749] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1b9105e3-b991-463a-94e1-903985e42e3e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.136627] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 879.136627] env[61911]: value = "task-1251193" [ 879.136627] env[61911]: _type = "Task" [ 879.136627] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.152122] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251193, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.251748] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a031e5-c4be-0307-a546-9005992299ec, 'name': SearchDatastore_Task, 'duration_secs': 0.020738} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.252583] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4495b7a9-14da-45e7-a54b-c52aa4449509 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.260238] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 879.260238] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523d95dc-bd6b-388d-58b0-aabdbdcad91a" [ 879.260238] env[61911]: _type = "Task" [ 879.260238] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.268153] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523d95dc-bd6b-388d-58b0-aabdbdcad91a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.408050] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.408050] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72c155a4-5cb8-45ff-b4e3-c00da583c25b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.417149] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 879.417149] env[61911]: value = "task-1251194" [ 879.417149] env[61911]: _type = "Task" [ 879.417149] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.427460] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.443278] env[61911]: DEBUG nova.compute.utils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.444826] env[61911]: DEBUG nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 879.445751] env[61911]: DEBUG nova.network.neutron [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 879.471456] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.471727] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.473103] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c86b3b-fde9-44f2-9ade-969673359386 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.494539] env[61911]: DEBUG nova.policy [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2266afee766b4660b9be910cd0c080c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f9c20b59cfc4a8389da83b0e6a87730', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.496737] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230c3ac1-3d16-446c-bbc4-b901d6f10efb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.525307] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Reconfiguring VM to detach interface {{(pid=61911) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 879.526033] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee56c4e5-3d3d-4cf3-852b-a66590e30810 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.545700] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 879.545700] env[61911]: value = "task-1251195" [ 879.545700] env[61911]: _type = "Task" [ 879.545700] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.555159] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.598855] env[61911]: DEBUG nova.network.neutron [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Successfully updated port: 82208eb4-5834-4277-9616-85003ea23d27 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.648521] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251193, 'name': Destroy_Task, 'duration_secs': 0.343874} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.648921] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Destroyed the VM [ 879.649192] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 879.649478] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a91fd451-88c0-4c4e-b98f-d809cc418abf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.653334] env[61911]: INFO nova.compute.manager [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Rebuilding instance [ 879.657155] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 879.657155] env[61911]: value = "task-1251196" [ 879.657155] env[61911]: _type = "Task" [ 879.657155] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.666934] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251196, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.697991] env[61911]: DEBUG nova.compute.manager [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.699253] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171d959d-6ac1-44d5-8cfc-fa637098d650 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.771642] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523d95dc-bd6b-388d-58b0-aabdbdcad91a, 'name': SearchDatastore_Task, 'duration_secs': 0.018205} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.771982] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.772258] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 5b7d09e1-ee14-4001-8d73-14e763402670/5b7d09e1-ee14-4001-8d73-14e763402670.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.772544] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fc06cbe-b44e-48f2-a475-07766612039d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.781522] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 879.781522] env[61911]: value = "task-1251197" [ 879.781522] env[61911]: _type = "Task" [ 879.781522] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.790634] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251197, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.854279] env[61911]: DEBUG nova.compute.manager [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Received event network-changed-82208eb4-5834-4277-9616-85003ea23d27 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 879.854475] env[61911]: DEBUG nova.compute.manager [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Refreshing instance network info cache due to event network-changed-82208eb4-5834-4277-9616-85003ea23d27. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 879.854762] env[61911]: DEBUG oslo_concurrency.lockutils [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] Acquiring lock "refresh_cache-da6f7457-e603-415c-9578-024a7563d34f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.854909] env[61911]: DEBUG oslo_concurrency.lockutils [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] Acquired lock "refresh_cache-da6f7457-e603-415c-9578-024a7563d34f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.855089] env[61911]: DEBUG nova.network.neutron [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Refreshing network info cache for port 82208eb4-5834-4277-9616-85003ea23d27 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.868089] env[61911]: DEBUG nova.network.neutron [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Successfully created port: 7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.934694] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251194, 'name': PowerOffVM_Task, 'duration_secs': 0.274906} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.935012] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.935884] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7600168b-e232-4ff0-87e7-482e6aa42efd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.960582] env[61911]: DEBUG nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 879.966203] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7d486bbc-d644-47bb-8127-315f9ac3ca8c tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.030s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.966443] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.876s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.968171] env[61911]: INFO nova.compute.claims [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.971831] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4d64fe-6db8-4106-ac7c-36d150986b09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.012769] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.013433] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2352fa24-f620-4910-b0ce-84061f5863c1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.022852] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 880.022852] env[61911]: value = "task-1251198" [ 880.022852] env[61911]: _type = "Task" [ 880.022852] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.033505] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251198, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.060160] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.104511] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "refresh_cache-da6f7457-e603-415c-9578-024a7563d34f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.173479] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251196, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.299260] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251197, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.392097] env[61911]: DEBUG nova.network.neutron [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.502243] env[61911]: DEBUG nova.network.neutron [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.546682] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 880.546925] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.547189] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.547667] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.547667] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.548189] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "b8cd7073-c712-4470-8ece-d144e8ce275f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.548738] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.550412] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da5e799c-b2d8-42b2-ba63-e8b8cb5105ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.564252] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.575753] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.575880] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.576582] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c8246a8-9ee5-4e04-8a1e-d09f9a34a0ad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.583460] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 880.583460] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aefa8d-34e0-a1eb-01b5-07fee9f42382" [ 880.583460] env[61911]: _type = "Task" [ 880.583460] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.594122] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aefa8d-34e0-a1eb-01b5-07fee9f42382, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.668798] env[61911]: DEBUG oslo_vmware.api [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251196, 'name': RemoveSnapshot_Task, 'duration_secs': 0.943771} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.669099] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 880.669350] env[61911]: INFO nova.compute.manager [None req-0786a4bf-9a16-4736-93f0-fe407a12d2ac tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Took 15.47 seconds to snapshot the instance on the hypervisor. [ 880.716386] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.716564] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6646e878-d8d9-4869-b6b9-10e7ff8dd5fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.725544] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 880.725544] env[61911]: value = "task-1251199" [ 880.725544] env[61911]: _type = "Task" [ 880.725544] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.734324] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.794310] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251197, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597331} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.794769] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 5b7d09e1-ee14-4001-8d73-14e763402670/5b7d09e1-ee14-4001-8d73-14e763402670.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.795079] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.795494] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84a91bab-d2ac-42bb-b1e6-48bcc5b34561 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.803335] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 880.803335] env[61911]: value = "task-1251200" [ 880.803335] env[61911]: _type = "Task" [ 880.803335] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.811438] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.978121] env[61911]: DEBUG nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 881.004638] env[61911]: DEBUG oslo_concurrency.lockutils [req-b9a5e68b-cdb4-4995-94e4-9531c0515a41 req-aa3d3785-c7ec-43f3-9b5e-332b97389d1d service nova] Releasing lock "refresh_cache-da6f7457-e603-415c-9578-024a7563d34f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.005019] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquired lock "refresh_cache-da6f7457-e603-415c-9578-024a7563d34f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.005248] env[61911]: DEBUG nova.network.neutron [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.011654] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 881.012051] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.012240] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 881.012447] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.012613] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 881.012749] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 881.012958] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 881.013143] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 881.013510] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 881.013510] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 881.013646] env[61911]: DEBUG nova.virt.hardware [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 881.014746] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecef589c-8595-4343-a839-694231fa6818 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.027032] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767afdaa-29e3-464c-bfa7-2f4560633157 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.053576] env[61911]: DEBUG nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 881.064299] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.094891] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aefa8d-34e0-a1eb-01b5-07fee9f42382, 'name': SearchDatastore_Task, 'duration_secs': 0.010095} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.098077] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15d02fe0-ceb3-4a26-a452-bc3c32b332e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.104842] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 881.104842] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521ccd68-06ad-ef70-89d4-ebd489f8a3b0" [ 881.104842] env[61911]: _type = "Task" [ 881.104842] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.115967] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521ccd68-06ad-ef70-89d4-ebd489f8a3b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.236499] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251199, 'name': PowerOffVM_Task, 'duration_secs': 0.439532} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.239189] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.239444] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.240417] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce52412-885d-43c0-b4fa-4f9d010487eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.247786] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.248053] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-33b199a3-9cc2-4ce3-88f6-3d998c71992f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.302310] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a302382a-4a31-4de6-be05-91e1ead56f6c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.313771] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067884} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.315824] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.316932] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96d6993-0ba9-4be7-920b-6d218911b24b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.320613] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f811d7-cbae-4ca0-b930-c6e517e8acb4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.351998] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 5b7d09e1-ee14-4001-8d73-14e763402670/5b7d09e1-ee14-4001-8d73-14e763402670.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.383748] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d38655e3-ab1b-47f5-9b1e-50385437b36e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.400771] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1567ca6d-60a4-4549-b13b-bcbed1455b62 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.404086] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.404463] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.404530] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleting the datastore file [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.405163] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25990b04-6507-427a-a15d-19fb61a1c869 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.415370] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9728b687-745e-483a-8041-03593c412118 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.419642] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 881.419642] env[61911]: value = "task-1251202" [ 881.419642] env[61911]: _type = "Task" [ 881.419642] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.419928] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 881.419928] env[61911]: value = "task-1251203" [ 881.419928] env[61911]: _type = "Task" [ 881.419928] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.440020] env[61911]: DEBUG nova.compute.provider_tree [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.444109] env[61911]: DEBUG nova.network.neutron [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Successfully updated port: 7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.454913] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.455152] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.546449] env[61911]: DEBUG nova.network.neutron [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.562447] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.583479] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.616533] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521ccd68-06ad-ef70-89d4-ebd489f8a3b0, 'name': SearchDatastore_Task, 'duration_secs': 0.011133} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.617290] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.617290] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. {{(pid=61911) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 881.617487] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a43fa548-fac6-4b49-814c-6545bbf319d0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.624925] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 881.624925] env[61911]: value = "task-1251204" [ 881.624925] env[61911]: _type = "Task" [ 881.624925] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.633999] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251204, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.719490] env[61911]: DEBUG nova.network.neutron [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Updating instance_info_cache with network_info: [{"id": "82208eb4-5834-4277-9616-85003ea23d27", "address": "fa:16:3e:14:54:da", "network": {"id": "f683f5b0-fbe0-4ad2-9883-c7d9caf4e674", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1247505226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "687cce9c9abb4c928295cbe60ec1d078", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee43879-c0b2-47f7-80d0-2c86e3d6d8b5", "external-id": "nsx-vlan-transportzone-151", "segmentation_id": 151, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82208eb4-58", "ovs_interfaceid": "82208eb4-5834-4277-9616-85003ea23d27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.901656] env[61911]: DEBUG nova.compute.manager [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Received event network-vif-plugged-7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 881.902523] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] Acquiring lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.902523] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.902523] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.902850] env[61911]: DEBUG nova.compute.manager [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] No waiting events found dispatching network-vif-plugged-7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 881.902850] env[61911]: WARNING nova.compute.manager [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Received unexpected event network-vif-plugged-7ce08c90-28e1-4989-9616-05bc8d5ee616 for instance with vm_state building and task_state spawning. [ 881.902911] env[61911]: DEBUG nova.compute.manager [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Received event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 881.903078] env[61911]: DEBUG nova.compute.manager [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing instance network info cache due to event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 881.903270] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] Acquiring lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.903410] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] Acquired lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.903572] env[61911]: DEBUG nova.network.neutron [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing network info cache for port 7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.934345] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.474502} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.938490] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 881.938757] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 881.938946] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.942192] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251202, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.952310] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "b6b16408-7633-4f23-8f57-1193b1fb0384" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.952618] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.952873] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "b6b16408-7633-4f23-8f57-1193b1fb0384-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.953304] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.953368] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.955146] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.955950] env[61911]: DEBUG nova.scheduler.client.report [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 881.959391] env[61911]: INFO nova.compute.manager [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Terminating instance [ 882.065010] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.138940] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251204, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.222339] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Releasing lock "refresh_cache-da6f7457-e603-415c-9578-024a7563d34f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.222858] env[61911]: DEBUG nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Instance network_info: |[{"id": "82208eb4-5834-4277-9616-85003ea23d27", "address": "fa:16:3e:14:54:da", "network": {"id": "f683f5b0-fbe0-4ad2-9883-c7d9caf4e674", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1247505226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "687cce9c9abb4c928295cbe60ec1d078", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ee43879-c0b2-47f7-80d0-2c86e3d6d8b5", "external-id": "nsx-vlan-transportzone-151", "segmentation_id": 151, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82208eb4-58", "ovs_interfaceid": "82208eb4-5834-4277-9616-85003ea23d27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 882.223554] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:54:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ee43879-c0b2-47f7-80d0-2c86e3d6d8b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82208eb4-5834-4277-9616-85003ea23d27', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.232575] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Creating folder: Project (687cce9c9abb4c928295cbe60ec1d078). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.232983] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6b3b0cf-e205-4f0a-8c52-8ba9075fa28a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.248669] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Created folder: Project (687cce9c9abb4c928295cbe60ec1d078) in parent group-v269521. [ 882.248906] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Creating folder: Instances. Parent ref: group-v269612. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.249592] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8d5654a-8dda-44c6-8c6e-1b4cb78abb2b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.262251] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Created folder: Instances in parent group-v269612. [ 882.262561] env[61911]: DEBUG oslo.service.loopingcall [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.262783] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da6f7457-e603-415c-9578-024a7563d34f] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.263030] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c520b06a-59da-4c56-bb65-54e3860567d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.284065] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.284065] env[61911]: value = "task-1251207" [ 882.284065] env[61911]: _type = "Task" [ 882.284065] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.292705] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251207, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.431954] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251202, 'name': ReconfigVM_Task, 'duration_secs': 0.585765} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.432290] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 5b7d09e1-ee14-4001-8d73-14e763402670/5b7d09e1-ee14-4001-8d73-14e763402670.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.432932] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eab49cab-f7b9-465d-be59-f9324efe9837 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.435794] env[61911]: DEBUG nova.network.neutron [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.440350] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 882.440350] env[61911]: value = "task-1251208" [ 882.440350] env[61911]: _type = "Task" [ 882.440350] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.453469] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251208, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.462524] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.463252] env[61911]: DEBUG nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 882.466056] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.821s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.466305] env[61911]: DEBUG nova.objects.instance [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lazy-loading 'resources' on Instance uuid 9230cd85-960d-427c-b69c-d19ac2bcfe35 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.467749] env[61911]: DEBUG nova.compute.manager [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 882.467994] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.469121] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7fd4b7-919f-4dd4-a1ac-08cd9c12b647 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.478127] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.478605] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-133def08-1bd2-4355-b050-e8767de5a6a2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.491215] env[61911]: DEBUG oslo_vmware.api [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 882.491215] env[61911]: value = "task-1251209" [ 882.491215] env[61911]: _type = "Task" [ 882.491215] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.502728] env[61911]: DEBUG oslo_vmware.api [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.533708] env[61911]: DEBUG nova.network.neutron [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.562415] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.637493] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251204, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677644} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.637917] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. [ 882.638828] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2b2d96-cbc1-4f05-b3ec-3060105f4d32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.673502] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.673831] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d8fcb80-4ab6-443f-bd01-bf6ca65b6912 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.694846] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 882.694846] env[61911]: value = "task-1251210" [ 882.694846] env[61911]: _type = "Task" [ 882.694846] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.704707] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251210, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.795395] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251207, 'name': CreateVM_Task, 'duration_secs': 0.400837} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.795603] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da6f7457-e603-415c-9578-024a7563d34f] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.796343] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.796517] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.796841] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.797118] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45f9e479-0231-4afd-9aa2-c5f076aa5171 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.802841] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 882.802841] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c88203-1407-1268-548e-1e062774d212" [ 882.802841] env[61911]: _type = "Task" [ 882.802841] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.811469] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c88203-1407-1268-548e-1e062774d212, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.955310] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251208, 'name': Rename_Task, 'duration_secs': 0.204863} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.958450] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.958777] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7df61ce4-4e15-4aa0-9284-d0be62c8ef80 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.966660] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 882.966660] env[61911]: value = "task-1251211" [ 882.966660] env[61911]: _type = "Task" [ 882.966660] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.974392] env[61911]: DEBUG nova.compute.utils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.983848] env[61911]: DEBUG nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 882.984779] env[61911]: DEBUG nova.network.neutron [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 882.990807] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251211, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.002343] env[61911]: DEBUG oslo_vmware.api [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251209, 'name': PowerOffVM_Task, 'duration_secs': 0.221592} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.004669] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 883.004896] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.005075] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 883.005271] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.005545] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 883.005655] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 883.005860] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 883.006777] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 883.007041] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 883.007434] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 883.007434] env[61911]: DEBUG nova.virt.hardware [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 883.007729] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 883.007898] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 883.009079] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36777f8-efb5-40aa-a6e0-8abd800e7fa2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.013899] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59907f19-bb5b-4a3c-9c60-9bffc27f6084 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.022037] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbe7ca4-9518-4813-8e5f-285a1d0c365d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.041853] env[61911]: DEBUG oslo_concurrency.lockutils [req-5bcc0335-0b60-48d5-8ecc-ba7515bf281c req-39b3b693-e007-44d1-a0b7-27c2b1ab6d5f service nova] Releasing lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.041853] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:fb:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90825141-c63e-4853-8f27-6553cc0f3ecc', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.048788] env[61911]: DEBUG oslo.service.loopingcall [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.053156] env[61911]: DEBUG nova.policy [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b034026eede4f5ca4757b552f7dca51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5be8ec1d59a14d618906ec32d0e134e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.055356] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.055356] env[61911]: DEBUG nova.network.neutron [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.056896] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 883.056896] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1c6ad7d-a1ae-42dd-84cf-d381a87c36bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.083336] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.084843] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.084843] env[61911]: value = "task-1251213" [ 883.084843] env[61911]: _type = "Task" [ 883.084843] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.098458] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.098458] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.099012] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Deleting the datastore file [datastore2] b6b16408-7633-4f23-8f57-1193b1fb0384 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.102079] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7cdb583-b175-46e2-98e9-09a7a74ba974 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.104738] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251213, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.114845] env[61911]: DEBUG oslo_vmware.api [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 883.114845] env[61911]: value = "task-1251214" [ 883.114845] env[61911]: _type = "Task" [ 883.114845] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.127304] env[61911]: DEBUG oslo_vmware.api [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.209823] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251210, 'name': ReconfigVM_Task, 'duration_secs': 0.291072} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.210227] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.211128] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8990275e-0d9e-4ad0-af0f-363758218698 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.955358] env[61911]: DEBUG nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 883.965090] env[61911]: DEBUG nova.network.neutron [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Successfully created port: 4e7ff300-bdd2-43d6-9fe5-646ffb508e5f {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.983197] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb73f725-3e75-42b1-82af-1c6aadcd89d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.006754] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.015983] env[61911]: DEBUG oslo_vmware.api [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251211, 'name': PowerOnVM_Task, 'duration_secs': 0.551021} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.016273] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c88203-1407-1268-548e-1e062774d212, 'name': SearchDatastore_Task, 'duration_secs': 0.011468} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.021814] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.021814] env[61911]: INFO nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Took 9.90 seconds to spawn the instance on the hypervisor. [ 884.021814] env[61911]: DEBUG nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 884.021814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.021814] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.021814] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.022119] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.022119] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.029591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096b15ec-d8de-4f98-bfa7-4afc8c390512 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.030094] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8112583-9816-4cfa-89cd-b174d3c2390e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.032420] env[61911]: DEBUG oslo_vmware.api [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251214, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147432} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.032726] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 884.032726] env[61911]: value = "task-1251215" [ 884.032726] env[61911]: _type = "Task" [ 884.032726] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.035987] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.036209] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.036455] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.036561] env[61911]: INFO nova.compute.manager [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Took 1.57 seconds to destroy the instance on the hypervisor. [ 884.036836] env[61911]: DEBUG oslo.service.loopingcall [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.037358] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251213, 'name': CreateVM_Task, 'duration_secs': 0.410484} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.038272] env[61911]: DEBUG nova.compute.manager [-] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 884.038373] env[61911]: DEBUG nova.network.neutron [-] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.043015] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 884.047717] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.047895] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.048297] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.052774] env[61911]: DEBUG nova.network.neutron [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.058335] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e57bd4c-dc23-4dac-89e7-037e31cc744e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.059231] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.059459] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 884.064129] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9439b53-e4d6-4832-8527-c643195dd43d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.066745] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251215, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.074900] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 884.074900] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525a5f88-4e93-d56c-a7dd-f29c01fc5c4e" [ 884.074900] env[61911]: _type = "Task" [ 884.074900] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.075300] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 884.075300] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a7ac61-579f-f7ef-1fa6-f46854fcb9cc" [ 884.075300] env[61911]: _type = "Task" [ 884.075300] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.092115] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a7ac61-579f-f7ef-1fa6-f46854fcb9cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010734} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.092115] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525a5f88-4e93-d56c-a7dd-f29c01fc5c4e, 'name': SearchDatastore_Task, 'duration_secs': 0.010308} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.099739] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.099739] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.099739] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.100307] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cef2a0cb-511d-445c-90f2-affcfdf75af2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.107787] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 884.107787] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a02524-09d9-84c0-5fc7-35028280f83b" [ 884.107787] env[61911]: _type = "Task" [ 884.107787] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.119588] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a02524-09d9-84c0-5fc7-35028280f83b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.164346] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce03cb19-428d-4a36-a62f-bf4011be743d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.173590] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eacfe01-1b85-4c2e-8df8-c2de260cf792 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.206388] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec688b76-b0f1-42f4-930b-04c73737994e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.214883] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f15f019-7f1c-4a1a-a6d7-e48758434060 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.229051] env[61911]: DEBUG nova.compute.provider_tree [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.318121] env[61911]: DEBUG nova.network.neutron [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [{"id": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "address": "fa:16:3e:c5:e0:b6", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce08c90-28", "ovs_interfaceid": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.468593] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.547858] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251215, 'name': ReconfigVM_Task, 'duration_secs': 0.162029} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.548206] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.548555] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfc22364-9b51-4b9e-a9ad-f252008d2302 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.555853] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 884.555853] env[61911]: value = "task-1251216" [ 884.555853] env[61911]: _type = "Task" [ 884.555853] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.568260] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.572782] env[61911]: INFO nova.compute.manager [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Took 26.82 seconds to build instance. [ 884.618478] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a02524-09d9-84c0-5fc7-35028280f83b, 'name': SearchDatastore_Task, 'duration_secs': 0.011165} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.618761] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.619035] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] da6f7457-e603-415c-9578-024a7563d34f/da6f7457-e603-415c-9578-024a7563d34f.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 884.619329] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.619516] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.619752] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05c9984e-faf8-42b9-941a-3f13d5206033 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.621729] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b66be9ae-8269-4707-bc91-bada29737f37 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.631888] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 884.631888] env[61911]: value = "task-1251217" [ 884.631888] env[61911]: _type = "Task" [ 884.631888] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.636123] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.636318] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 884.637397] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92fb8603-88ca-4a5e-b7f1-f4fc2471d876 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.642830] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.644136] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 884.644136] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a121d5-c89e-b137-8ea8-2acd23da9c62" [ 884.644136] env[61911]: _type = "Task" [ 884.644136] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.652265] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a121d5-c89e-b137-8ea8-2acd23da9c62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.732440] env[61911]: DEBUG nova.scheduler.client.report [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 884.821101] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.821539] env[61911]: DEBUG nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Instance network_info: |[{"id": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "address": "fa:16:3e:c5:e0:b6", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce08c90-28", "ovs_interfaceid": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 884.822135] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:e0:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ce08c90-28e1-4989-9616-05bc8d5ee616', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.830599] env[61911]: DEBUG oslo.service.loopingcall [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.830875] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.831148] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1da8a5db-ab0b-4c38-a088-7126b8f0a3bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.848347] env[61911]: DEBUG nova.compute.manager [req-480347fd-df3c-4f08-8727-ecdecd1611da req-9f083d43-f16c-4836-83ff-35ad1c0d6831 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Received event network-vif-deleted-6e7b50c2-7b5e-4b94-8267-825879051466 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 884.848534] env[61911]: INFO nova.compute.manager [req-480347fd-df3c-4f08-8727-ecdecd1611da req-9f083d43-f16c-4836-83ff-35ad1c0d6831 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Neutron deleted interface 6e7b50c2-7b5e-4b94-8267-825879051466; detaching it from the instance and deleting it from the info cache [ 884.848825] env[61911]: DEBUG nova.network.neutron [req-480347fd-df3c-4f08-8727-ecdecd1611da req-9f083d43-f16c-4836-83ff-35ad1c0d6831 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.855846] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.855846] env[61911]: value = "task-1251218" [ 884.855846] env[61911]: _type = "Task" [ 884.855846] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.866269] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251218, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.972061] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.975226] env[61911]: DEBUG nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 885.017816] env[61911]: DEBUG nova.virt.hardware [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 885.017816] env[61911]: DEBUG nova.network.neutron [-] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.019340] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c80b353-8b3f-42d4-b2ba-d1d6323c59a6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.031215] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15fd85e1-d84d-4846-b3a6-3564c291b1a6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.067375] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251216, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.077236] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23535feb-5129-4333-974d-4e079c8d1ef0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.342s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.147784] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251217, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.159248] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a121d5-c89e-b137-8ea8-2acd23da9c62, 'name': SearchDatastore_Task, 'duration_secs': 0.009552} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.160330] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-819b73e4-bdd7-46db-8f22-053a6a946e2d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.167944] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 885.167944] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52025262-9dc1-5873-dc8d-05816e0d6252" [ 885.167944] env[61911]: _type = "Task" [ 885.167944] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.178042] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52025262-9dc1-5873-dc8d-05816e0d6252, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.238234] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.772s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.240556] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.316s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.243531] env[61911]: INFO nova.compute.claims [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.277950] env[61911]: INFO nova.scheduler.client.report [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted allocations for instance 9230cd85-960d-427c-b69c-d19ac2bcfe35 [ 885.351611] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec9ee72d-eab6-402f-801d-f914c97d0846 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.365769] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a80a06-6dde-4f26-bf24-6a8e7b13ee4d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.384210] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251218, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.409459] env[61911]: DEBUG nova.compute.manager [req-480347fd-df3c-4f08-8727-ecdecd1611da req-9f083d43-f16c-4836-83ff-35ad1c0d6831 service nova] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Detach interface failed, port_id=6e7b50c2-7b5e-4b94-8267-825879051466, reason: Instance b6b16408-7633-4f23-8f57-1193b1fb0384 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 885.469167] env[61911]: DEBUG oslo_vmware.api [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251195, 'name': ReconfigVM_Task, 'duration_secs': 5.793703} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.469370] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.469741] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Reconfigured VM to detach interface {{(pid=61911) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 885.525815] env[61911]: INFO nova.compute.manager [-] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Took 1.49 seconds to deallocate network for instance. [ 885.568555] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251216, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.644771] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590501} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.646151] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] da6f7457-e603-415c-9578-024a7563d34f/da6f7457-e603-415c-9578-024a7563d34f.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 885.646151] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.646151] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31e180ca-7615-4208-886f-24ed91f44a6b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.653991] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 885.653991] env[61911]: value = "task-1251219" [ 885.653991] env[61911]: _type = "Task" [ 885.653991] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.663542] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251219, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.678573] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52025262-9dc1-5873-dc8d-05816e0d6252, 'name': SearchDatastore_Task, 'duration_secs': 0.037101} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.678944] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.679333] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 885.679881] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-468c2ea6-c74b-453b-9519-00731afbfe66 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.684378] env[61911]: DEBUG nova.compute.manager [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Received event network-changed-5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 885.684598] env[61911]: DEBUG nova.compute.manager [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Refreshing instance network info cache due to event network-changed-5c7df9be-477d-4aab-9d5b-e340866dc6bb. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 885.684825] env[61911]: DEBUG oslo_concurrency.lockutils [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] Acquiring lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.684987] env[61911]: DEBUG oslo_concurrency.lockutils [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] Acquired lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.685170] env[61911]: DEBUG nova.network.neutron [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Refreshing network info cache for port 5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.688091] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 885.688091] env[61911]: value = "task-1251220" [ 885.688091] env[61911]: _type = "Task" [ 885.688091] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.699571] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.790202] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b845c343-e47e-4ac7-8c83-2244868a90be tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "9230cd85-960d-427c-b69c-d19ac2bcfe35" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.638s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.873715] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251218, 'name': CreateVM_Task, 'duration_secs': 0.636449} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.874237] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.875317] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.875626] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.876703] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.878055] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-093a62e5-584d-4895-a8ba-3292ce4ef10d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.886785] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 885.886785] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5289ad03-e84a-f0c9-3206-29e235e90a82" [ 885.886785] env[61911]: _type = "Task" [ 885.886785] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.899350] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5289ad03-e84a-f0c9-3206-29e235e90a82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.921561] env[61911]: DEBUG nova.network.neutron [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Successfully updated port: 4e7ff300-bdd2-43d6-9fe5-646ffb508e5f {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.037955] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.067918] env[61911]: DEBUG oslo_vmware.api [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251216, 'name': PowerOnVM_Task, 'duration_secs': 1.230243} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.068261] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.071260] env[61911]: DEBUG nova.compute.manager [None req-94257c98-b896-4b1c-8978-a9465b7295d3 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.072107] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5b16d5-1cf8-458e-addf-bf31a28ef9f5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.166561] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251219, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.129378} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.170018] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.170018] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b08f87-5fb7-4add-8002-165e6524ef82 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.191705] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] da6f7457-e603-415c-9578-024a7563d34f/da6f7457-e603-415c-9578-024a7563d34f.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.194686] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acf49d40-33ab-4af1-8160-98dc9bb852f9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.221577] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524801} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.222035] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.223217] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.223861] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf8f6986-2169-4db8-ac8b-2ef99aff877a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.226251] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 886.226251] env[61911]: value = "task-1251221" [ 886.226251] env[61911]: _type = "Task" [ 886.226251] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.232472] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 886.232472] env[61911]: value = "task-1251222" [ 886.232472] env[61911]: _type = "Task" [ 886.232472] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.239474] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251221, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.248411] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251222, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.401152] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5289ad03-e84a-f0c9-3206-29e235e90a82, 'name': SearchDatastore_Task, 'duration_secs': 0.053382} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.401495] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.401946] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.402071] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.402264] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.402452] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.402915] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-920c9838-f99f-4d4c-b886-0522beb92a62 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.414460] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.414704] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.415542] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17820028-ed01-478a-992b-5b061870f0ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.427020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.427020] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.427020] env[61911]: DEBUG nova.network.neutron [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.427020] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 886.427020] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b28e9-8729-30a2-c376-f9f9eabec702" [ 886.427020] env[61911]: _type = "Task" [ 886.427020] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.436869] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b28e9-8729-30a2-c376-f9f9eabec702, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.580691] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9126ae60-23a4-4654-b778-d683ea82e09a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.597096] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2360f6c-8fdb-4a0b-8af5-40debb8bbfd9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.637573] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc2af35-1b34-4d07-8168-d6a5221e9c79 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.646664] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c427e4-c339-4bb9-8914-04a844a81a61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.666534] env[61911]: DEBUG nova.compute.provider_tree [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.677283] env[61911]: DEBUG nova.network.neutron [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updated VIF entry in instance network info cache for port 5c7df9be-477d-4aab-9d5b-e340866dc6bb. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.677722] env[61911]: DEBUG nova.network.neutron [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updating instance_info_cache with network_info: [{"id": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "address": "fa:16:3e:39:28:bf", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c7df9be-47", "ovs_interfaceid": "5c7df9be-477d-4aab-9d5b-e340866dc6bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.736949] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.747512] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251222, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117745} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.748020] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.749023] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a7f307-c9c2-4376-9301-45a0fdf513bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.776195] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.776540] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81ccca12-1249-499f-8fdb-434fd937f433 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.797377] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 886.797377] env[61911]: value = "task-1251223" [ 886.797377] env[61911]: _type = "Task" [ 886.797377] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.806610] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251223, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.941383] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b28e9-8729-30a2-c376-f9f9eabec702, 'name': SearchDatastore_Task, 'duration_secs': 0.019386} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.942211] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a04aceb9-a464-4642-9404-7f1226c39ee9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.949670] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 886.949670] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52833e28-356e-1f38-45b7-5e8ebedb271f" [ 886.949670] env[61911]: _type = "Task" [ 886.949670] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.961044] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52833e28-356e-1f38-45b7-5e8ebedb271f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.975808] env[61911]: DEBUG nova.network.neutron [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.097422] env[61911]: INFO nova.compute.manager [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Unrescuing [ 887.097792] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.097928] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquired lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.098118] env[61911]: DEBUG nova.network.neutron [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.153485] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.153714] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.153878] env[61911]: DEBUG nova.network.neutron [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.169466] env[61911]: DEBUG nova.scheduler.client.report [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 887.180634] env[61911]: DEBUG oslo_concurrency.lockutils [req-cb6846f3-2b21-464f-8a06-166bfd4f955c req-1dcf2c55-b1e2-4816-8c7e-f827ed1d9481 service nova] Releasing lock "refresh_cache-5b7d09e1-ee14-4001-8d73-14e763402670" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.242023] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251221, 'name': ReconfigVM_Task, 'duration_secs': 0.992303} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.242023] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Reconfigured VM instance instance-0000004c to attach disk [datastore1] da6f7457-e603-415c-9578-024a7563d34f/da6f7457-e603-415c-9578-024a7563d34f.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.242023] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1589431-4925-4352-bd7f-4b5cb718c7a7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.248479] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 887.248479] env[61911]: value = "task-1251224" [ 887.248479] env[61911]: _type = "Task" [ 887.248479] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.257105] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251224, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.258219] env[61911]: DEBUG nova.network.neutron [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance_info_cache with network_info: [{"id": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "address": "fa:16:3e:c2:34:3f", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7ff300-bd", "ovs_interfaceid": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.308786] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251223, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.462215] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52833e28-356e-1f38-45b7-5e8ebedb271f, 'name': SearchDatastore_Task, 'duration_secs': 0.011431} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.462509] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.462783] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 887.463063] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ceac1ca7-5b13-47c8-80c4-9a2b2da69351 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.470553] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 887.470553] env[61911]: value = "task-1251225" [ 887.470553] env[61911]: _type = "Task" [ 887.470553] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.479132] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251225, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.675764] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.676448] env[61911]: DEBUG nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 887.679663] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.637s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.756421] env[61911]: DEBUG nova.compute.manager [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Received event network-vif-plugged-4e7ff300-bdd2-43d6-9fe5-646ffb508e5f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 887.756636] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] Acquiring lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.756822] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.757012] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.757199] env[61911]: DEBUG nova.compute.manager [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] No waiting events found dispatching network-vif-plugged-4e7ff300-bdd2-43d6-9fe5-646ffb508e5f {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 887.757767] env[61911]: WARNING nova.compute.manager [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Received unexpected event network-vif-plugged-4e7ff300-bdd2-43d6-9fe5-646ffb508e5f for instance with vm_state building and task_state spawning. [ 887.757767] env[61911]: DEBUG nova.compute.manager [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Received event network-changed-4e7ff300-bdd2-43d6-9fe5-646ffb508e5f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 887.757767] env[61911]: DEBUG nova.compute.manager [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Refreshing instance network info cache due to event network-changed-4e7ff300-bdd2-43d6-9fe5-646ffb508e5f. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 887.757915] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] Acquiring lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.763139] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.763139] env[61911]: DEBUG nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Instance network_info: |[{"id": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "address": "fa:16:3e:c2:34:3f", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7ff300-bd", "ovs_interfaceid": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 887.763385] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251224, 'name': Rename_Task, 'duration_secs': 0.327627} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.765874] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] Acquired lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.766052] env[61911]: DEBUG nova.network.neutron [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Refreshing network info cache for port 4e7ff300-bdd2-43d6-9fe5-646ffb508e5f {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.769755] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:34:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e7ff300-bdd2-43d6-9fe5-646ffb508e5f', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.779370] env[61911]: DEBUG oslo.service.loopingcall [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.780729] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 887.784173] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.784173] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c87c84bb-cbd8-4ab2-99ec-6c5f370aa011 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.790115] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81c6f0ac-470f-4b9a-a4f5-17b440c110e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.815982] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 887.815982] env[61911]: value = "task-1251226" [ 887.815982] env[61911]: _type = "Task" [ 887.815982] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.823052] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251223, 'name': ReconfigVM_Task, 'duration_secs': 0.906664} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.827810] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.827810] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.827810] env[61911]: value = "task-1251227" [ 887.827810] env[61911]: _type = "Task" [ 887.827810] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.828734] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87ebe995-1b1c-4382-b814-99d63bca0067 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.838967] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251226, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.846176] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251227, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.846624] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 887.846624] env[61911]: value = "task-1251228" [ 887.846624] env[61911]: _type = "Task" [ 887.846624] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.861878] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251228, 'name': Rename_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.985305] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251225, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.107552] env[61911]: DEBUG nova.network.neutron [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Updating instance_info_cache with network_info: [{"id": "0e16550a-783a-473b-9be5-cb569b373191", "address": "fa:16:3e:a7:ad:f8", "network": {"id": "2f20ed58-2f99-4dc1-ac1c-bd3aa46c7b36", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-819531908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ada275c0bead46f8a9d295952cd930ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e16550a-78", "ovs_interfaceid": "0e16550a-783a-473b-9be5-cb569b373191", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.185431] env[61911]: DEBUG nova.compute.utils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.193023] env[61911]: DEBUG nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 888.193023] env[61911]: DEBUG nova.network.neutron [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 888.207908] env[61911]: DEBUG nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 888.259159] env[61911]: INFO nova.network.neutron [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Port c34b4a10-4268-487c-bb19-1ef7beda1f06 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 888.259159] env[61911]: DEBUG nova.network.neutron [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.302126] env[61911]: DEBUG nova.policy [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe50c877fcef489cb798fafe73caaf8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813b3181120245e594809d096a621675', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.328922] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251226, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.345910] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251227, 'name': CreateVM_Task, 'duration_secs': 0.44738} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.346128] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.346991] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.347326] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.347727] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.348651] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cd30de5-99a9-48d0-80ff-e905790fcc60 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.361074] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 888.361074] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526b2f0b-4aa7-5abd-dbf1-eabe15cdcbdb" [ 888.361074] env[61911]: _type = "Task" [ 888.361074] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.364204] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251228, 'name': Rename_Task, 'duration_secs': 0.229117} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.368065] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.368360] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6f2e163-e6d0-4517-8d01-8d4ca6e83bc3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.378676] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526b2f0b-4aa7-5abd-dbf1-eabe15cdcbdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.378676] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 888.378676] env[61911]: value = "task-1251229" [ 888.378676] env[61911]: _type = "Task" [ 888.378676] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.379161] env[61911]: DEBUG nova.network.neutron [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updated VIF entry in instance network info cache for port 4e7ff300-bdd2-43d6-9fe5-646ffb508e5f. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.379828] env[61911]: DEBUG nova.network.neutron [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance_info_cache with network_info: [{"id": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "address": "fa:16:3e:c2:34:3f", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7ff300-bd", "ovs_interfaceid": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.391600] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.485352] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251225, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546123} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.485352] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 888.485352] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.485352] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e4d2880-59ef-4d38-8ff2-91f086b81f32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.492935] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 888.492935] env[61911]: value = "task-1251230" [ 888.492935] env[61911]: _type = "Task" [ 888.492935] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.504496] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251230, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.610991] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Releasing lock "refresh_cache-1058f8d2-0979-4423-802d-52d91d5a2ff7" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.611657] env[61911]: DEBUG nova.objects.instance [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lazy-loading 'flavor' on Instance uuid 1058f8d2-0979-4423-802d-52d91d5a2ff7 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.743654] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 66472f43-537d-4eb3-8d49-d40627a8809d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.743816] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 267075d7-c3d8-4520-b7d9-67504097f823 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.743941] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 28159e75-9fe9-44c7-b5c9-534902cecbef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744140] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8051e8ec-6d82-416c-858a-f0cdd00f869d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744208] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 31f0a4f7-a69d-4ec4-966d-93f402163b88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744304] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance e9833b95-4162-42ba-87a4-d4cc790ac8e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744416] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bafd58bc-8ae7-49a4-a039-31d328699010 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744528] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance c95707ca-7e32-42c0-914a-f0b178f07997 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744638] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 83616524-2530-4398-9019-148e0658720d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744748] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 613544bc-92b9-4c8f-8a7b-f02205f347ff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744856] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance f0dcfbbf-98ab-4e9f-838f-d9601d2bb045 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.744958] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance e3ad3416-423c-4687-aa35-2fb2ef67ec64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.745102] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 9c2e75d8-b307-43bb-84a8-604cd5edbc39 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.745191] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 1058f8d2-0979-4423-802d-52d91d5a2ff7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.745631] env[61911]: WARNING nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b6b16408-7633-4f23-8f57-1193b1fb0384 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 888.745631] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5b7d09e1-ee14-4001-8d73-14e763402670 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.745631] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance da6f7457-e603-415c-9578-024a7563d34f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.745631] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.745812] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance a35a11f1-3d4b-439e-9517-fa5a9f43af60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.745851] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 50a522ab-651a-4bb3-93d3-8866169e95e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.761162] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.818089] env[61911]: DEBUG nova.network.neutron [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Successfully created port: 76870f3b-7732-4724-8dc0-38e3012a2fee {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.830951] env[61911]: DEBUG oslo_vmware.api [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251226, 'name': PowerOnVM_Task, 'duration_secs': 0.595855} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.831418] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.831505] env[61911]: INFO nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Took 10.35 seconds to spawn the instance on the hypervisor. [ 888.831695] env[61911]: DEBUG nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 888.832520] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2654e0d8-ec4f-49d4-b2f1-12cf34fdbb7c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.876634] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526b2f0b-4aa7-5abd-dbf1-eabe15cdcbdb, 'name': SearchDatastore_Task, 'duration_secs': 0.027685} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.878070] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.878341] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.878896] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.879105] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.879303] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.880040] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de701e84-2aaa-43a5-ae4d-f5f214fa3daa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.886562] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b778604-c82a-4538-923a-b552de91567f req-c5b861cc-ac7e-4d1c-9ad7-feaf0ef04be6 service nova] Releasing lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.894565] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251229, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.901358] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.901553] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.902294] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-880e26c2-cda8-45cd-b070-092a6f5c8238 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.908939] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 888.908939] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac8a2e-8e6b-75eb-450e-9d90762e7d84" [ 888.908939] env[61911]: _type = "Task" [ 888.908939] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.918055] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac8a2e-8e6b-75eb-450e-9d90762e7d84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.002738] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251230, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.178224} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.003066] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.003796] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f34183a-0653-45bb-aaba-eede83a6fb2a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.026391] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.026694] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73631afd-fc07-4094-b5bb-88d7ee9248a2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.050308] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 889.050308] env[61911]: value = "task-1251231" [ 889.050308] env[61911]: _type = "Task" [ 889.050308] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.060763] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251231, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.120518] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48db646-527f-499d-a4d8-17adf9fae255 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.143440] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.143814] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b268e1b-eabd-4990-b754-5b2ea700641f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.152114] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 889.152114] env[61911]: value = "task-1251232" [ 889.152114] env[61911]: _type = "Task" [ 889.152114] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.161823] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251232, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.227937] env[61911]: DEBUG nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 889.230655] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "interface-e3ad3416-423c-4687-aa35-2fb2ef67ec64-c34b4a10-4268-487c-bb19-1ef7beda1f06" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.230787] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e3ad3416-423c-4687-aa35-2fb2ef67ec64-c34b4a10-4268-487c-bb19-1ef7beda1f06" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.231386] env[61911]: DEBUG nova.objects.instance [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lazy-loading 'flavor' on Instance uuid e3ad3416-423c-4687-aa35-2fb2ef67ec64 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.249864] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b8cd7073-c712-4470-8ece-d144e8ce275f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 889.249864] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 889.249864] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4160MB phys_disk=200GB used_disk=19GB total_vcpus=48 used_vcpus=19 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 889.263022] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 889.263022] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.263022] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 889.263278] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.263278] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 889.263377] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 889.263597] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 889.263785] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 889.264014] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 889.264198] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 889.264379] env[61911]: DEBUG nova.virt.hardware [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 889.265541] env[61911]: DEBUG oslo_concurrency.lockutils [None req-185b57a8-7c14-4276-98bd-417e81bb92d3 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e9833b95-4162-42ba-87a4-d4cc790ac8e5-c34b4a10-4268-487c-bb19-1ef7beda1f06" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.298s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.267031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3519950-0de8-4e6c-9db0-07cbaadd1c52 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.281155] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cd8546-94a8-4a3c-88a1-53d2cdee09cf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.357289] env[61911]: INFO nova.compute.manager [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Took 24.85 seconds to build instance. [ 889.396409] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251229, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.424575] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ac8a2e-8e6b-75eb-450e-9d90762e7d84, 'name': SearchDatastore_Task, 'duration_secs': 0.035071} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.425551] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84fcf1c0-dbe4-488c-8fdf-2027e1354966 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.435121] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 889.435121] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524e94c4-0be9-a5d8-ce24-5e2f156cecc4" [ 889.435121] env[61911]: _type = "Task" [ 889.435121] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.446711] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524e94c4-0be9-a5d8-ce24-5e2f156cecc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.565756] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251231, 'name': ReconfigVM_Task, 'duration_secs': 0.276405} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.566373] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.567466] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a6485a01-213f-4929-b18a-a179cc98dc2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.570430] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f25c657-d276-4fed-860e-1c0cc86eb284 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.574051] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 889.574051] env[61911]: value = "task-1251233" [ 889.574051] env[61911]: _type = "Task" [ 889.574051] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.580705] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24df59c-c120-4b44-88d4-4b1e8f652794 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.587265] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251233, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.619086] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396044be-7193-4ae4-b9d9-60720df1bf61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.628196] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07f774a-f068-43b1-b3a5-a2d23d04e0bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.644782] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.664473] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251232, 'name': PowerOffVM_Task, 'duration_secs': 0.496065} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.665209] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.670440] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 889.671075] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de6d7e74-3d03-4e09-880d-b6219c77adb5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.692692] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 889.692692] env[61911]: value = "task-1251234" [ 889.692692] env[61911]: _type = "Task" [ 889.692692] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.702191] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251234, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.838016] env[61911]: DEBUG nova.objects.instance [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lazy-loading 'pci_requests' on Instance uuid e3ad3416-423c-4687-aa35-2fb2ef67ec64 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.857593] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4c48008-c220-4c2c-a5ef-cc95bc88485f tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "da6f7457-e603-415c-9578-024a7563d34f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.358s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.896466] env[61911]: DEBUG oslo_vmware.api [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251229, 'name': PowerOnVM_Task, 'duration_secs': 1.187476} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.896715] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.896948] env[61911]: DEBUG nova.compute.manager [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 889.897852] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9bd6ff-4a96-4f6c-9d6f-90906928d3f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.948363] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524e94c4-0be9-a5d8-ce24-5e2f156cecc4, 'name': SearchDatastore_Task, 'duration_secs': 0.010139} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.951019] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.951019] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] a35a11f1-3d4b-439e-9517-fa5a9f43af60/a35a11f1-3d4b-439e-9517-fa5a9f43af60.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.951019] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8de7042-6d7e-4fea-af5a-7ccde9d2c9ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.959200] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 889.959200] env[61911]: value = "task-1251235" [ 889.959200] env[61911]: _type = "Task" [ 889.959200] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.968427] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251235, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.088319] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251233, 'name': Rename_Task, 'duration_secs': 0.152824} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.088640] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.088929] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66639479-0869-4ef1-93f0-979a46c98ff9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.096493] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 890.096493] env[61911]: value = "task-1251236" [ 890.096493] env[61911]: _type = "Task" [ 890.096493] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.107909] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.148323] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 890.207798] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251234, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.268028] env[61911]: DEBUG nova.compute.manager [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 890.268028] env[61911]: DEBUG nova.compute.manager [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing instance network info cache due to event network-changed-c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 890.268028] env[61911]: DEBUG oslo_concurrency.lockutils [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] Acquiring lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.268028] env[61911]: DEBUG oslo_concurrency.lockutils [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] Acquired lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.268028] env[61911]: DEBUG nova.network.neutron [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Refreshing network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.340253] env[61911]: DEBUG nova.objects.base [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 890.340574] env[61911]: DEBUG nova.network.neutron [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 890.415890] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.417792] env[61911]: DEBUG nova.policy [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3ee96b52414148b7bff1d067595848b1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f40f512384d4fedb75bfc53f3195243', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 890.469550] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251235, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480028} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.469833] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] a35a11f1-3d4b-439e-9517-fa5a9f43af60/a35a11f1-3d4b-439e-9517-fa5a9f43af60.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.470097] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.470378] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85898d57-6265-4c8f-96bf-b0a3fdf10bc6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.479592] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 890.479592] env[61911]: value = "task-1251237" [ 890.479592] env[61911]: _type = "Task" [ 890.479592] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.488805] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251237, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.557066] env[61911]: DEBUG nova.network.neutron [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Successfully updated port: 76870f3b-7732-4724-8dc0-38e3012a2fee {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.611045] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251236, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.653598] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 890.653936] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.974s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.654251] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.071s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.655942] env[61911]: INFO nova.compute.claims [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.707292] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251234, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.716686] env[61911]: DEBUG nova.compute.manager [req-a0324a78-f1c9-41f7-839d-5f520e709efb req-cfcd3e42-8fe8-430c-9702-d96464dc0d87 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Received event network-vif-plugged-76870f3b-7732-4724-8dc0-38e3012a2fee {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 890.717018] env[61911]: DEBUG oslo_concurrency.lockutils [req-a0324a78-f1c9-41f7-839d-5f520e709efb req-cfcd3e42-8fe8-430c-9702-d96464dc0d87 service nova] Acquiring lock "50a522ab-651a-4bb3-93d3-8866169e95e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.717349] env[61911]: DEBUG oslo_concurrency.lockutils [req-a0324a78-f1c9-41f7-839d-5f520e709efb req-cfcd3e42-8fe8-430c-9702-d96464dc0d87 service nova] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.717619] env[61911]: DEBUG oslo_concurrency.lockutils [req-a0324a78-f1c9-41f7-839d-5f520e709efb req-cfcd3e42-8fe8-430c-9702-d96464dc0d87 service nova] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.717945] env[61911]: DEBUG nova.compute.manager [req-a0324a78-f1c9-41f7-839d-5f520e709efb req-cfcd3e42-8fe8-430c-9702-d96464dc0d87 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] No waiting events found dispatching network-vif-plugged-76870f3b-7732-4724-8dc0-38e3012a2fee {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 890.718236] env[61911]: WARNING nova.compute.manager [req-a0324a78-f1c9-41f7-839d-5f520e709efb req-cfcd3e42-8fe8-430c-9702-d96464dc0d87 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Received unexpected event network-vif-plugged-76870f3b-7732-4724-8dc0-38e3012a2fee for instance with vm_state building and task_state spawning. [ 890.841075] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "da6f7457-e603-415c-9578-024a7563d34f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.841421] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "da6f7457-e603-415c-9578-024a7563d34f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.841567] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "da6f7457-e603-415c-9578-024a7563d34f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.841758] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "da6f7457-e603-415c-9578-024a7563d34f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.841939] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "da6f7457-e603-415c-9578-024a7563d34f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.844168] env[61911]: INFO nova.compute.manager [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Terminating instance [ 890.993503] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251237, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.064813] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "refresh_cache-50a522ab-651a-4bb3-93d3-8866169e95e4" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.064813] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "refresh_cache-50a522ab-651a-4bb3-93d3-8866169e95e4" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.064813] env[61911]: DEBUG nova.network.neutron [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.112024] env[61911]: DEBUG oslo_vmware.api [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251236, 'name': PowerOnVM_Task, 'duration_secs': 0.692949} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.112024] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.112024] env[61911]: INFO nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Took 10.13 seconds to spawn the instance on the hypervisor. [ 891.112024] env[61911]: DEBUG nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 891.112024] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f79dc4c-f2d6-4679-99b2-9c2b5577e098 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.206148] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251234, 'name': ReconfigVM_Task, 'duration_secs': 1.438541} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.207936] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 891.208369] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.208771] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ed7f6c8-3c99-4a5a-baf8-72ea2f9c10f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.217092] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 891.217092] env[61911]: value = "task-1251238" [ 891.217092] env[61911]: _type = "Task" [ 891.217092] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.226216] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251238, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.286377] env[61911]: DEBUG nova.network.neutron [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updated VIF entry in instance network info cache for port c0501390-f2c7-4b6d-bd50-5d3a03c0e130. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 891.286377] env[61911]: DEBUG nova.network.neutron [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [{"id": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "address": "fa:16:3e:38:9f:ea", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0501390-f2", "ovs_interfaceid": "c0501390-f2c7-4b6d-bd50-5d3a03c0e130", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.351027] env[61911]: DEBUG nova.compute.manager [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 891.351027] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.351027] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16360f1-3c44-4cba-bac8-8c3af0a07ba2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.362417] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.363077] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9a46230-5a26-4cda-a6cc-afdc79b33577 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.372411] env[61911]: DEBUG oslo_vmware.api [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 891.372411] env[61911]: value = "task-1251239" [ 891.372411] env[61911]: _type = "Task" [ 891.372411] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.382522] env[61911]: DEBUG oslo_vmware.api [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251239, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.491953] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251237, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.773508} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.492309] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.493170] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02470af-02ce-45a1-ae42-2201f67fa998 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.517352] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] a35a11f1-3d4b-439e-9517-fa5a9f43af60/a35a11f1-3d4b-439e-9517-fa5a9f43af60.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.517665] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc18fb2b-3797-4aa6-9e55-02fe075c990a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.538414] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 891.538414] env[61911]: value = "task-1251240" [ 891.538414] env[61911]: _type = "Task" [ 891.538414] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.549113] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251240, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.600357] env[61911]: DEBUG nova.network.neutron [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.629254] env[61911]: INFO nova.compute.manager [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Took 26.75 seconds to build instance. [ 891.733870] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251238, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.789209] env[61911]: DEBUG oslo_concurrency.lockutils [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] Releasing lock "refresh_cache-e9833b95-4162-42ba-87a4-d4cc790ac8e5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.789421] env[61911]: DEBUG nova.compute.manager [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 891.789603] env[61911]: DEBUG nova.compute.manager [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing instance network info cache due to event network-changed-de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 891.789862] env[61911]: DEBUG oslo_concurrency.lockutils [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.790038] env[61911]: DEBUG oslo_concurrency.lockutils [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.790213] env[61911]: DEBUG nova.network.neutron [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.832344] env[61911]: DEBUG nova.network.neutron [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Updating instance_info_cache with network_info: [{"id": "76870f3b-7732-4724-8dc0-38e3012a2fee", "address": "fa:16:3e:36:9d:c9", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76870f3b-77", "ovs_interfaceid": "76870f3b-7732-4724-8dc0-38e3012a2fee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.885344] env[61911]: DEBUG oslo_vmware.api [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251239, 'name': PowerOffVM_Task, 'duration_secs': 0.314893} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.885652] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.885859] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.886096] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06c5bcc8-886d-40de-af64-5a2adccb0e12 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.950364] env[61911]: DEBUG nova.network.neutron [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Successfully updated port: c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.969674] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.969925] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.970142] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Deleting the datastore file [datastore1] da6f7457-e603-415c-9578-024a7563d34f {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.970413] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15f78841-6212-443b-9a16-3b7aacf4680f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.980487] env[61911]: DEBUG oslo_vmware.api [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for the task: (returnval){ [ 891.980487] env[61911]: value = "task-1251242" [ 891.980487] env[61911]: _type = "Task" [ 891.980487] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.994023] env[61911]: DEBUG oslo_vmware.api [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.001756] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca2381b-23ba-47cf-9745-5252082636ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.010769] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d810da-cf5c-4caa-acc2-de28dc3e8cca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.048776] env[61911]: INFO nova.compute.manager [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Rebuilding instance [ 892.054296] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c226eba3-ea1e-4a6d-97f8-ee3e6dda8f75 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.067649] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251240, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.068808] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d010cb4c-99b2-4ef2-a638-bd4ccf51fbb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.087846] env[61911]: DEBUG nova.compute.provider_tree [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.112320] env[61911]: DEBUG nova.compute.manager [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 892.113239] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0eb6b0-74ad-4dc7-a23d-74575f867cec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.133331] env[61911]: DEBUG oslo_concurrency.lockutils [None req-db0afa5a-2555-4819-b785-64f23376be08 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.262s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.229818] env[61911]: DEBUG oslo_vmware.api [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251238, 'name': PowerOnVM_Task, 'duration_secs': 0.530555} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.230210] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.230507] env[61911]: DEBUG nova.compute.manager [None req-9adadf46-85c6-4986-a21a-0dbfa02d95f2 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 892.231367] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d600ae-2746-4d57-bade-45cec96e3276 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.341100] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "refresh_cache-50a522ab-651a-4bb3-93d3-8866169e95e4" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.341436] env[61911]: DEBUG nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Instance network_info: |[{"id": "76870f3b-7732-4724-8dc0-38e3012a2fee", "address": "fa:16:3e:36:9d:c9", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76870f3b-77", "ovs_interfaceid": "76870f3b-7732-4724-8dc0-38e3012a2fee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 892.341897] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:9d:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76870f3b-7732-4724-8dc0-38e3012a2fee', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.349558] env[61911]: DEBUG oslo.service.loopingcall [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.349793] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 892.350032] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5cdc44e6-6214-4165-8617-38c95d327b0c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.370772] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.370772] env[61911]: value = "task-1251243" [ 892.370772] env[61911]: _type = "Task" [ 892.370772] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.381252] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251243, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.452776] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.495880] env[61911]: DEBUG oslo_vmware.api [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Task: {'id': task-1251242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209927} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.495880] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.496260] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.496260] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.497510] env[61911]: INFO nova.compute.manager [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] [instance: da6f7457-e603-415c-9578-024a7563d34f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 892.497510] env[61911]: DEBUG oslo.service.loopingcall [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.497510] env[61911]: DEBUG nova.compute.manager [-] [instance: da6f7457-e603-415c-9578-024a7563d34f] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 892.497510] env[61911]: DEBUG nova.network.neutron [-] [instance: da6f7457-e603-415c-9578-024a7563d34f] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 892.541294] env[61911]: DEBUG nova.network.neutron [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updated VIF entry in instance network info cache for port de067cc9-827d-4e43-a53e-6eda53bf9d1e. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.541680] env[61911]: DEBUG nova.network.neutron [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.563220] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251240, 'name': ReconfigVM_Task, 'duration_secs': 0.627168} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.563220] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Reconfigured VM instance instance-0000004e to attach disk [datastore1] a35a11f1-3d4b-439e-9517-fa5a9f43af60/a35a11f1-3d4b-439e-9517-fa5a9f43af60.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.563220] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-511284ac-d44a-4419-b139-eb116a202e40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.573086] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 892.573086] env[61911]: value = "task-1251244" [ 892.573086] env[61911]: _type = "Task" [ 892.573086] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.589385] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251244, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.595011] env[61911]: DEBUG nova.scheduler.client.report [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 892.881582] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251243, 'name': CreateVM_Task, 'duration_secs': 0.421702} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.881781] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.882486] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.882659] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.883609] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.883880] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbf5641d-2f8b-4c22-91f6-38f9bb63f069 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.889319] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 892.889319] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5220bfb9-9311-1e98-31e0-5867f85588b2" [ 892.889319] env[61911]: _type = "Task" [ 892.889319] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.897828] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5220bfb9-9311-1e98-31e0-5867f85588b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.951914] env[61911]: DEBUG nova.compute.manager [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-vif-plugged-c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 892.952186] env[61911]: DEBUG oslo_concurrency.lockutils [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] Acquiring lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.952383] env[61911]: DEBUG oslo_concurrency.lockutils [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.952533] env[61911]: DEBUG oslo_concurrency.lockutils [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.952834] env[61911]: DEBUG nova.compute.manager [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] No waiting events found dispatching network-vif-plugged-c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 892.953093] env[61911]: WARNING nova.compute.manager [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received unexpected event network-vif-plugged-c34b4a10-4268-487c-bb19-1ef7beda1f06 for instance with vm_state active and task_state None. [ 892.953272] env[61911]: DEBUG nova.compute.manager [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-changed-c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 892.953434] env[61911]: DEBUG nova.compute.manager [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing instance network info cache due to event network-changed-c34b4a10-4268-487c-bb19-1ef7beda1f06. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 892.953606] env[61911]: DEBUG oslo_concurrency.lockutils [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.044388] env[61911]: DEBUG oslo_concurrency.lockutils [req-a48317f0-1ffd-41af-85d8-72ed93cac20b req-77445416-3208-46a9-8d82-dafbc05aaa84 service nova] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.046228] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.046228] env[61911]: DEBUG nova.network.neutron [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.086125] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251244, 'name': Rename_Task, 'duration_secs': 0.199887} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.086586] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.086947] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-725178a1-9d63-43c8-bb2a-d102e196d6d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.097555] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 893.097555] env[61911]: value = "task-1251245" [ 893.097555] env[61911]: _type = "Task" [ 893.097555] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.103491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.104254] env[61911]: DEBUG nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 893.108443] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.071s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.108763] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.111610] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.696s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.111907] env[61911]: DEBUG nova.objects.instance [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 893.124561] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251245, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.130305] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 893.130305] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-560fa1dd-aee8-4556-9121-b7629ece7196 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.138933] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 893.138933] env[61911]: value = "task-1251246" [ 893.138933] env[61911]: _type = "Task" [ 893.138933] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.153663] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251246, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.166425] env[61911]: INFO nova.scheduler.client.report [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Deleted allocations for instance b6b16408-7633-4f23-8f57-1193b1fb0384 [ 893.189224] env[61911]: DEBUG nova.compute.manager [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Received event network-changed-76870f3b-7732-4724-8dc0-38e3012a2fee {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 893.189477] env[61911]: DEBUG nova.compute.manager [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Refreshing instance network info cache due to event network-changed-76870f3b-7732-4724-8dc0-38e3012a2fee. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 893.189712] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] Acquiring lock "refresh_cache-50a522ab-651a-4bb3-93d3-8866169e95e4" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.189875] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] Acquired lock "refresh_cache-50a522ab-651a-4bb3-93d3-8866169e95e4" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.190141] env[61911]: DEBUG nova.network.neutron [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Refreshing network info cache for port 76870f3b-7732-4724-8dc0-38e3012a2fee {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 893.404690] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5220bfb9-9311-1e98-31e0-5867f85588b2, 'name': SearchDatastore_Task, 'duration_secs': 0.034974} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.405163] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.405302] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.405556] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.405766] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.405915] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.406223] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a7e46b0-038d-4f79-b2c6-211f1c745431 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.416277] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.416478] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 893.417235] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7132184-5849-45b6-bc79-f6c58c2ea76c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.422994] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 893.422994] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c52082-81e1-f9f5-676f-b1c99c73569a" [ 893.422994] env[61911]: _type = "Task" [ 893.422994] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.431488] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c52082-81e1-f9f5-676f-b1c99c73569a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.537327] env[61911]: DEBUG nova.network.neutron [-] [instance: da6f7457-e603-415c-9578-024a7563d34f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.606097] env[61911]: WARNING nova.network.neutron [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] 87858d49-056a-4528-87e1-d07b16c7f333 already exists in list: networks containing: ['87858d49-056a-4528-87e1-d07b16c7f333']. ignoring it [ 893.618267] env[61911]: DEBUG nova.compute.utils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.620211] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251245, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.626328] env[61911]: DEBUG nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 893.626615] env[61911]: DEBUG nova.network.neutron [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 893.655165] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251246, 'name': PowerOffVM_Task, 'duration_secs': 0.276952} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.655621] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.655992] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 893.657262] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8eff79-b1e8-410e-b943-b840e2a2b9bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.669571] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 893.675303] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9e14162-040c-4878-afb0-3ea438d0036e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.682914] env[61911]: DEBUG nova.policy [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d0056b50ef7429ea460ff4c1156d132', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf7a7083777a43a5bb9fc04584dcdd06', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.685495] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1d46e913-a444-46c7-8f7e-c9305399bbfb tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "b6b16408-7633-4f23-8f57-1193b1fb0384" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.732s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.754486] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "1058f8d2-0979-4423-802d-52d91d5a2ff7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.754733] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.754941] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "1058f8d2-0979-4423-802d-52d91d5a2ff7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.760030] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.760307] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.005s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.762525] env[61911]: INFO nova.compute.manager [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Terminating instance [ 893.791255] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 893.791585] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 893.791837] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleting the datastore file [datastore1] 66472f43-537d-4eb3-8d49-d40627a8809d {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.792571] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-962a1f5d-7faf-475e-aa46-af99acc4c70f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.801208] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 893.801208] env[61911]: value = "task-1251249" [ 893.801208] env[61911]: _type = "Task" [ 893.801208] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.812440] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251249, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.935405] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c52082-81e1-f9f5-676f-b1c99c73569a, 'name': SearchDatastore_Task, 'duration_secs': 0.012468} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.936685] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f55c0425-33aa-40ef-8cba-ecc7e28d9e67 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.943710] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 893.943710] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526ff327-b57d-63d3-fe16-93ea853c45f5" [ 893.943710] env[61911]: _type = "Task" [ 893.943710] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.955590] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526ff327-b57d-63d3-fe16-93ea853c45f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.040095] env[61911]: INFO nova.compute.manager [-] [instance: da6f7457-e603-415c-9578-024a7563d34f] Took 1.54 seconds to deallocate network for instance. [ 894.057647] env[61911]: DEBUG nova.network.neutron [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Updated VIF entry in instance network info cache for port 76870f3b-7732-4724-8dc0-38e3012a2fee. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 894.058104] env[61911]: DEBUG nova.network.neutron [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Updating instance_info_cache with network_info: [{"id": "76870f3b-7732-4724-8dc0-38e3012a2fee", "address": "fa:16:3e:36:9d:c9", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76870f3b-77", "ovs_interfaceid": "76870f3b-7732-4724-8dc0-38e3012a2fee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.110466] env[61911]: DEBUG oslo_vmware.api [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251245, 'name': PowerOnVM_Task, 'duration_secs': 0.588577} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.110765] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.110979] env[61911]: INFO nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Took 9.14 seconds to spawn the instance on the hypervisor. [ 894.111198] env[61911]: DEBUG nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 894.111991] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423d0a55-f91c-4306-9c0b-2e56ec338994 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.125025] env[61911]: DEBUG nova.network.neutron [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Successfully created port: 3a287a5b-935a-49ef-b44f-2e53162f4440 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.125025] env[61911]: DEBUG nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 894.129187] env[61911]: DEBUG nova.network.neutron [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "address": "fa:16:3e:b6:3a:b4", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc34b4a10-42", "ovs_interfaceid": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.131260] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9173c6a2-513c-468a-b4ab-a001a29d787b tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.266571] env[61911]: DEBUG nova.compute.manager [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 894.266709] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.268111] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef7caad-7b0b-4f1a-a6dd-e5f73bb696fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.277758] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.278088] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e05ec02a-8b9e-4f61-8854-d2f9ace82695 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.285558] env[61911]: DEBUG oslo_vmware.api [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 894.285558] env[61911]: value = "task-1251250" [ 894.285558] env[61911]: _type = "Task" [ 894.285558] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.294821] env[61911]: DEBUG oslo_vmware.api [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.311229] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251249, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257135} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.311520] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 894.311798] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 894.311926] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 894.456331] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526ff327-b57d-63d3-fe16-93ea853c45f5, 'name': SearchDatastore_Task, 'duration_secs': 0.040291} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.456686] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.456853] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 50a522ab-651a-4bb3-93d3-8866169e95e4/50a522ab-651a-4bb3-93d3-8866169e95e4.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 894.457146] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e2c9e5e-7b6e-453e-a22c-fab7bccaf609 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.465265] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 894.465265] env[61911]: value = "task-1251251" [ 894.465265] env[61911]: _type = "Task" [ 894.465265] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.474110] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.551037] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.551037] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.551037] env[61911]: DEBUG nova.objects.instance [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lazy-loading 'resources' on Instance uuid da6f7457-e603-415c-9578-024a7563d34f {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.564024] env[61911]: DEBUG oslo_concurrency.lockutils [req-5b86f661-2968-457d-8202-d2e9f48a7c36 req-29298b45-5fd6-4003-8e4a-21f927e49c08 service nova] Releasing lock "refresh_cache-50a522ab-651a-4bb3-93d3-8866169e95e4" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.636432] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.636432] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.636978] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.638807] env[61911]: DEBUG oslo_concurrency.lockutils [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.639012] env[61911]: DEBUG nova.network.neutron [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Refreshing network info cache for port c34b4a10-4268-487c-bb19-1ef7beda1f06 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 894.642548] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a318530-9c39-4bf7-b01a-130e4069618e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.644026] env[61911]: INFO nova.compute.manager [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Took 20.57 seconds to build instance. [ 894.664932] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 894.665182] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.665382] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 894.665607] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.665789] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 894.665982] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 894.666260] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 894.666471] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 894.666753] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 894.666962] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 894.667378] env[61911]: DEBUG nova.virt.hardware [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 894.674542] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Reconfiguring VM to attach interface {{(pid=61911) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 894.675665] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ca66f69-fd66-4107-aeb2-7a343231f305 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.696059] env[61911]: DEBUG oslo_vmware.api [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 894.696059] env[61911]: value = "task-1251252" [ 894.696059] env[61911]: _type = "Task" [ 894.696059] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.705285] env[61911]: DEBUG oslo_vmware.api [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251252, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.796338] env[61911]: DEBUG oslo_vmware.api [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251250, 'name': PowerOffVM_Task, 'duration_secs': 0.283552} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.796632] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.796877] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.797156] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-04d87e60-6de5-4d2e-b527-09323203498c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.891665] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 894.892052] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 894.892298] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Deleting the datastore file [datastore2] 1058f8d2-0979-4423-802d-52d91d5a2ff7 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.892589] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8321c299-85b5-44b8-9f18-a03f457a6dae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.900773] env[61911]: DEBUG oslo_vmware.api [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 894.900773] env[61911]: value = "task-1251254" [ 894.900773] env[61911]: _type = "Task" [ 894.900773] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.911799] env[61911]: DEBUG oslo_vmware.api [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251254, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.976668] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251251, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.138713] env[61911]: DEBUG nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 895.147279] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0b3bea15-b727-4dcf-beff-617f28339510 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.086s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.171031] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b3c56e0598a511f250557794f65f3085',container_format='bare',created_at=2024-10-10T15:59:37Z,direct_url=,disk_format='vmdk',id=37eeba24-1178-4ae0-aa71-cb4902ea42c8,min_disk=1,min_ram=0,name='tempest-test-snap-405511292',owner='cf7a7083777a43a5bb9fc04584dcdd06',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T15:59:52Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 895.171198] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.171249] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 895.171448] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.171597] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 895.171748] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 895.171983] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 895.172159] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 895.172336] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 895.172510] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 895.172692] env[61911]: DEBUG nova.virt.hardware [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 895.173655] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a500dc51-a952-48d0-8726-cdffb31b1330 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.187331] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c0c7fc-ff61-4872-942f-e6dbab9c685e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.216859] env[61911]: DEBUG oslo_vmware.api [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251252, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.344757] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 895.344997] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.345274] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 895.345741] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.345741] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 895.345867] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 895.346032] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 895.346266] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 895.346492] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 895.346688] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 895.346911] env[61911]: DEBUG nova.virt.hardware [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 895.347890] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1636f7c-bd61-436f-9318-aea8108855e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.353427] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e299f31d-ad01-4036-a4f0-fa9fdc4ec9d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.361403] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545e5505-2805-43ff-b4d6-53b715e6e514 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.371487] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-307e6a32-108c-40cd-9fd2-a754c0188c7d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.388494] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:fb:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90825141-c63e-4853-8f27-6553cc0f3ecc', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.397226] env[61911]: DEBUG oslo.service.loopingcall [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.398714] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 895.398959] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5b39b38-c81d-414a-98cb-44dfa230b440 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.442535] env[61911]: DEBUG nova.network.neutron [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updated VIF entry in instance network info cache for port c34b4a10-4268-487c-bb19-1ef7beda1f06. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.442535] env[61911]: DEBUG nova.network.neutron [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "address": "fa:16:3e:b6:3a:b4", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc34b4a10-42", "ovs_interfaceid": "c34b4a10-4268-487c-bb19-1ef7beda1f06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.449151] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e80f46c-0055-4edc-8bb4-312f45c3c038 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.459847] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.459847] env[61911]: value = "task-1251255" [ 895.459847] env[61911]: _type = "Task" [ 895.459847] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.460356] env[61911]: DEBUG oslo_vmware.api [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251254, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.46161} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.460616] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.460892] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.461300] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.461400] env[61911]: INFO nova.compute.manager [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Took 1.19 seconds to destroy the instance on the hypervisor. [ 895.461634] env[61911]: DEBUG oslo.service.loopingcall [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.463151] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fcdbf7-563c-4ceb-8cc8-bfa59c86b060 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.471407] env[61911]: DEBUG nova.compute.manager [-] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 895.471551] env[61911]: DEBUG nova.network.neutron [-] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 895.491830] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251255, 'name': CreateVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.492744] env[61911]: DEBUG nova.compute.provider_tree [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.496776] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.810631} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.497272] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 50a522ab-651a-4bb3-93d3-8866169e95e4/50a522ab-651a-4bb3-93d3-8866169e95e4.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 895.497506] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.497757] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23a79925-822b-4dcf-aa3a-777c7896d6a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.505158] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 895.505158] env[61911]: value = "task-1251256" [ 895.505158] env[61911]: _type = "Task" [ 895.505158] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.517635] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251256, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.709397] env[61911]: DEBUG oslo_vmware.api [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251252, 'name': ReconfigVM_Task, 'duration_secs': 0.963775} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.710603] env[61911]: DEBUG nova.network.neutron [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Successfully updated port: 3a287a5b-935a-49ef-b44f-2e53162f4440 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.712468] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.712722] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Reconfigured VM to attach interface {{(pid=61911) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 895.899161] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.899161] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.899399] env[61911]: INFO nova.compute.manager [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Shelving [ 895.905928] env[61911]: DEBUG nova.compute.manager [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] [instance: da6f7457-e603-415c-9578-024a7563d34f] Received event network-vif-deleted-82208eb4-5834-4277-9616-85003ea23d27 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 895.905979] env[61911]: DEBUG nova.compute.manager [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Received event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 895.906195] env[61911]: DEBUG nova.compute.manager [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing instance network info cache due to event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 895.906466] env[61911]: DEBUG oslo_concurrency.lockutils [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] Acquiring lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.906736] env[61911]: DEBUG oslo_concurrency.lockutils [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] Acquired lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.906861] env[61911]: DEBUG nova.network.neutron [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing network info cache for port 7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 895.951441] env[61911]: DEBUG oslo_concurrency.lockutils [req-1c3932ca-bdae-4916-8af0-011b3351c422 req-6c510da6-74ed-464a-a248-b3c515551574 service nova] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.970205] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251255, 'name': CreateVM_Task, 'duration_secs': 0.483952} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.970434] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.971121] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.971305] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.971626] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.971883] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-858f42b3-5e94-460c-b6f7-ca25de7e6503 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.977383] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 895.977383] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528d06b3-517f-ae92-bdf0-62bda67c286c" [ 895.977383] env[61911]: _type = "Task" [ 895.977383] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.986312] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528d06b3-517f-ae92-bdf0-62bda67c286c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.998352] env[61911]: DEBUG nova.scheduler.client.report [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 896.015439] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251256, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071376} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.015703] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.016535] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d42694-b8fa-4c33-a097-971dde32fa05 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.038962] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 50a522ab-651a-4bb3-93d3-8866169e95e4/50a522ab-651a-4bb3-93d3-8866169e95e4.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.041604] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5baf6b9-0af9-428c-a50f-a0691f08e316 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.062273] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 896.062273] env[61911]: value = "task-1251257" [ 896.062273] env[61911]: _type = "Task" [ 896.062273] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.070758] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251257, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.216492] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "refresh_cache-b8cd7073-c712-4470-8ece-d144e8ce275f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.216492] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "refresh_cache-b8cd7073-c712-4470-8ece-d144e8ce275f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.216682] env[61911]: DEBUG nova.network.neutron [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.219080] env[61911]: DEBUG oslo_concurrency.lockutils [None req-cb6e34ef-4895-4b85-ad60-03c8a4dc84ac tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e3ad3416-423c-4687-aa35-2fb2ef67ec64-c34b4a10-4268-487c-bb19-1ef7beda1f06" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.988s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.281945] env[61911]: DEBUG nova.network.neutron [-] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.290081] env[61911]: DEBUG nova.compute.manager [req-a2056dab-fbb8-4710-84b3-79e517c45cae req-f04fa57a-1118-446d-a275-30d750e70aa6 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Received event network-vif-plugged-3a287a5b-935a-49ef-b44f-2e53162f4440 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 896.290419] env[61911]: DEBUG oslo_concurrency.lockutils [req-a2056dab-fbb8-4710-84b3-79e517c45cae req-f04fa57a-1118-446d-a275-30d750e70aa6 service nova] Acquiring lock "b8cd7073-c712-4470-8ece-d144e8ce275f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.290636] env[61911]: DEBUG oslo_concurrency.lockutils [req-a2056dab-fbb8-4710-84b3-79e517c45cae req-f04fa57a-1118-446d-a275-30d750e70aa6 service nova] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.290860] env[61911]: DEBUG oslo_concurrency.lockutils [req-a2056dab-fbb8-4710-84b3-79e517c45cae req-f04fa57a-1118-446d-a275-30d750e70aa6 service nova] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.291138] env[61911]: DEBUG nova.compute.manager [req-a2056dab-fbb8-4710-84b3-79e517c45cae req-f04fa57a-1118-446d-a275-30d750e70aa6 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] No waiting events found dispatching network-vif-plugged-3a287a5b-935a-49ef-b44f-2e53162f4440 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 896.291301] env[61911]: WARNING nova.compute.manager [req-a2056dab-fbb8-4710-84b3-79e517c45cae req-f04fa57a-1118-446d-a275-30d750e70aa6 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Received unexpected event network-vif-plugged-3a287a5b-935a-49ef-b44f-2e53162f4440 for instance with vm_state building and task_state spawning. [ 896.488962] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528d06b3-517f-ae92-bdf0-62bda67c286c, 'name': SearchDatastore_Task, 'duration_secs': 0.050316} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.489326] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.489662] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.489779] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.489943] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.490142] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.490416] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d514035a-53e2-43b8-b5d0-e68074acff59 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.503274] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.952s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.506409] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.506600] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.507342] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8d87a98-8562-451b-a435-8e0adca202b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.513364] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 896.513364] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5298d233-8663-6f57-f980-175df228945c" [ 896.513364] env[61911]: _type = "Task" [ 896.513364] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.521710] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5298d233-8663-6f57-f980-175df228945c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.530804] env[61911]: INFO nova.scheduler.client.report [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Deleted allocations for instance da6f7457-e603-415c-9578-024a7563d34f [ 896.575539] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251257, 'name': ReconfigVM_Task, 'duration_secs': 0.286175} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.576126] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 50a522ab-651a-4bb3-93d3-8866169e95e4/50a522ab-651a-4bb3-93d3-8866169e95e4.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.576601] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77b3bd9a-8627-4676-8ca4-14f978cb224f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.585625] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 896.585625] env[61911]: value = "task-1251258" [ 896.585625] env[61911]: _type = "Task" [ 896.585625] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.596160] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251258, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.659618] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "28159e75-9fe9-44c7-b5c9-534902cecbef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.659877] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.660126] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "28159e75-9fe9-44c7-b5c9-534902cecbef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.660327] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.660503] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.663228] env[61911]: INFO nova.compute.manager [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Terminating instance [ 896.725956] env[61911]: DEBUG nova.network.neutron [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updated VIF entry in instance network info cache for port 7ce08c90-28e1-4989-9616-05bc8d5ee616. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 896.726477] env[61911]: DEBUG nova.network.neutron [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [{"id": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "address": "fa:16:3e:c5:e0:b6", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce08c90-28", "ovs_interfaceid": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.754405] env[61911]: DEBUG nova.network.neutron [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 896.785651] env[61911]: INFO nova.compute.manager [-] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Took 1.31 seconds to deallocate network for instance. [ 896.894331] env[61911]: DEBUG nova.network.neutron [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Updating instance_info_cache with network_info: [{"id": "3a287a5b-935a-49ef-b44f-2e53162f4440", "address": "fa:16:3e:2d:d2:36", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a287a5b-93", "ovs_interfaceid": "3a287a5b-935a-49ef-b44f-2e53162f4440", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.909413] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.911188] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64cbae0c-cac8-48b7-970f-8b8a45b32c05 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.922019] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 896.922019] env[61911]: value = "task-1251259" [ 896.922019] env[61911]: _type = "Task" [ 896.922019] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.938487] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251259, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.027393] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5298d233-8663-6f57-f980-175df228945c, 'name': SearchDatastore_Task, 'duration_secs': 0.018701} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.027393] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b69adba-3d28-4bb8-9add-f608d817383f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.033101] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 897.033101] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f83d99-d0d8-5218-f57d-d0f7c9ef20a3" [ 897.033101] env[61911]: _type = "Task" [ 897.033101] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.042462] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f83d99-d0d8-5218-f57d-d0f7c9ef20a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.043162] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7921ec19-f70b-400a-be85-8fcc35c7e306 tempest-ServerPasswordTestJSON-1982213132 tempest-ServerPasswordTestJSON-1982213132-project-member] Lock "da6f7457-e603-415c-9578-024a7563d34f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.202s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.097557] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251258, 'name': Rename_Task, 'duration_secs': 0.158315} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.097858] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.098159] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7071d70b-5202-4ccf-b9ab-92ec66962b86 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.107946] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 897.107946] env[61911]: value = "task-1251260" [ 897.107946] env[61911]: _type = "Task" [ 897.107946] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.118237] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251260, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.168755] env[61911]: DEBUG nova.compute.manager [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 897.168995] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.169950] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e6ffe2-c81b-4154-bab8-0393301cf7ed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.180586] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.180821] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6cb0c551-b16e-4fa0-89a6-5383c667e54d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.188690] env[61911]: DEBUG oslo_vmware.api [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 897.188690] env[61911]: value = "task-1251261" [ 897.188690] env[61911]: _type = "Task" [ 897.188690] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.199090] env[61911]: DEBUG oslo_vmware.api [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251261, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.229803] env[61911]: DEBUG oslo_concurrency.lockutils [req-3e9547a6-6b37-4d9f-ac9d-db1cf0530351 req-9eb1057d-ff2a-4cba-9edf-314edcc9c2d2 service nova] Releasing lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.293621] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.294067] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.294374] env[61911]: DEBUG nova.objects.instance [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lazy-loading 'resources' on Instance uuid 1058f8d2-0979-4423-802d-52d91d5a2ff7 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.399094] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "refresh_cache-b8cd7073-c712-4470-8ece-d144e8ce275f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.399825] env[61911]: DEBUG nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Instance network_info: |[{"id": "3a287a5b-935a-49ef-b44f-2e53162f4440", "address": "fa:16:3e:2d:d2:36", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a287a5b-93", "ovs_interfaceid": "3a287a5b-935a-49ef-b44f-2e53162f4440", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 897.401920] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:d2:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a287a5b-935a-49ef-b44f-2e53162f4440', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.409933] env[61911]: DEBUG oslo.service.loopingcall [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.410324] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.410687] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7d57af1-ce1c-4214-8d2d-bc235ba0dd35 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.439183] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251259, 'name': PowerOffVM_Task, 'duration_secs': 0.455694} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.440593] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.440690] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.440690] env[61911]: value = "task-1251262" [ 897.440690] env[61911]: _type = "Task" [ 897.440690] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.441620] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cee1ad-2607-4d90-aa75-9b290574ce1a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.457134] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251262, 'name': CreateVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.471422] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e23713-9b3a-4096-bd58-73ef205406af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.546125] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f83d99-d0d8-5218-f57d-d0f7c9ef20a3, 'name': SearchDatastore_Task, 'duration_secs': 0.014683} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.546490] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.546722] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.547012] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d308036-d967-45df-9105-1ba1c7f827fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.555364] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 897.555364] env[61911]: value = "task-1251263" [ 897.555364] env[61911]: _type = "Task" [ 897.555364] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.566374] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251263, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.569201] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "interface-e3ad3416-423c-4687-aa35-2fb2ef67ec64-c34b4a10-4268-487c-bb19-1ef7beda1f06" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.569656] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e3ad3416-423c-4687-aa35-2fb2ef67ec64-c34b4a10-4268-487c-bb19-1ef7beda1f06" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.621220] env[61911]: DEBUG oslo_vmware.api [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251260, 'name': PowerOnVM_Task, 'duration_secs': 0.488925} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.621220] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.621507] env[61911]: INFO nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Took 8.39 seconds to spawn the instance on the hypervisor. [ 897.621761] env[61911]: DEBUG nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 897.622929] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7171bbb1-ad08-4d6b-bc34-5645e55d2bb9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.700146] env[61911]: DEBUG oslo_vmware.api [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251261, 'name': PowerOffVM_Task, 'duration_secs': 0.210342} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.700606] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.700851] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.701167] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67303236-b506-4418-9bd6-ed4887d923ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.806242] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.806242] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.806242] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Deleting the datastore file [datastore2] 28159e75-9fe9-44c7-b5c9-534902cecbef {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.806242] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1918029f-8ae1-4d7c-a33a-02f66782b335 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.812860] env[61911]: DEBUG oslo_vmware.api [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for the task: (returnval){ [ 897.812860] env[61911]: value = "task-1251265" [ 897.812860] env[61911]: _type = "Task" [ 897.812860] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.827152] env[61911]: DEBUG oslo_vmware.api [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251265, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.938031] env[61911]: DEBUG nova.compute.manager [req-87c6c4c4-8d69-4390-9cae-a25bf4a60fa9 req-346c9bd6-4be8-4aee-884d-77060cf962ed service nova] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Received event network-vif-deleted-0e16550a-783a-473b-9be5-cb569b373191 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 897.957094] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251262, 'name': CreateVM_Task, 'duration_secs': 0.392516} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.957289] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.958073] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.958309] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.958687] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.958965] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e6310e0-ec2e-4455-83c6-799ac6199b18 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.966384] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 897.966384] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52227bfc-6af2-646e-1086-4ac575c06494" [ 897.966384] env[61911]: _type = "Task" [ 897.966384] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.982625] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52227bfc-6af2-646e-1086-4ac575c06494, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.983984] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 897.984295] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bffc3f6d-9c28-4d9a-a4d6-d3de7f9d4c8b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.994442] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 897.994442] env[61911]: value = "task-1251266" [ 897.994442] env[61911]: _type = "Task" [ 897.994442] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.005938] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251266, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.068779] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251263, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.073076] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.073409] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.074318] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cda0d73-fd9b-4c09-9795-4be4218e0cbb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.103808] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847c3e1f-e6df-4182-a91b-e6339feb09b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.134823] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Reconfiguring VM to detach interface {{(pid=61911) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 898.142229] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20b2cd31-63fc-4661-9d5c-ea2326493a3e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.158417] env[61911]: INFO nova.compute.manager [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Took 19.26 seconds to build instance. [ 898.164618] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 898.164618] env[61911]: value = "task-1251267" [ 898.164618] env[61911]: _type = "Task" [ 898.164618] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.176657] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.186329] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4da06a-6a37-4d74-afa5-c0fbcf84a8b4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.195320] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f76ad5a-a627-48ab-90a1-a9ed98784df6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.227085] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917c42ac-1776-4611-908a-3e79eac4f04a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.237583] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9000040-b169-460e-bf1d-825d491221ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.252048] env[61911]: DEBUG nova.compute.provider_tree [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.324034] env[61911]: DEBUG oslo_vmware.api [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Task: {'id': task-1251265, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.460439} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.325041] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.325280] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.325529] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.325750] env[61911]: INFO nova.compute.manager [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Took 1.16 seconds to destroy the instance on the hypervisor. [ 898.326018] env[61911]: DEBUG oslo.service.loopingcall [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.326245] env[61911]: DEBUG nova.compute.manager [-] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 898.326350] env[61911]: DEBUG nova.network.neutron [-] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.449027] env[61911]: DEBUG nova.compute.manager [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Stashing vm_state: active {{(pid=61911) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 898.482452] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.482452] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Processing image 37eeba24-1178-4ae0-aa71-cb4902ea42c8 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.482452] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8/37eeba24-1178-4ae0-aa71-cb4902ea42c8.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.482452] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8/37eeba24-1178-4ae0-aa71-cb4902ea42c8.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.482452] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.482829] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f49839c-1533-4663-813f-be51e606d9f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.491770] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.491960] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.492702] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52ce1488-3850-4722-856b-493902a440ff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.504012] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251266, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.505329] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 898.505329] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5296f579-81c4-b236-6d20-b48e32f52349" [ 898.505329] env[61911]: _type = "Task" [ 898.505329] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.513802] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5296f579-81c4-b236-6d20-b48e32f52349, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.567947] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251263, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604021} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.569018] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.569018] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.569018] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c023cf3-4633-47ec-a924-22c91802fedd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.576473] env[61911]: DEBUG nova.compute.manager [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Received event network-changed-3a287a5b-935a-49ef-b44f-2e53162f4440 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 898.576787] env[61911]: DEBUG nova.compute.manager [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Refreshing instance network info cache due to event network-changed-3a287a5b-935a-49ef-b44f-2e53162f4440. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 898.579020] env[61911]: DEBUG oslo_concurrency.lockutils [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] Acquiring lock "refresh_cache-b8cd7073-c712-4470-8ece-d144e8ce275f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.579020] env[61911]: DEBUG oslo_concurrency.lockutils [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] Acquired lock "refresh_cache-b8cd7073-c712-4470-8ece-d144e8ce275f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.579020] env[61911]: DEBUG nova.network.neutron [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Refreshing network info cache for port 3a287a5b-935a-49ef-b44f-2e53162f4440 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.580854] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 898.580854] env[61911]: value = "task-1251268" [ 898.580854] env[61911]: _type = "Task" [ 898.580854] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.593318] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251268, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.662995] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a84adbe-4b22-4927-a0bf-26e806463263 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.774s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.676210] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.755074] env[61911]: DEBUG nova.scheduler.client.report [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 898.981746] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.009139] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251266, 'name': CreateSnapshot_Task, 'duration_secs': 0.756216} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.012094] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 899.015367] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32c1383-0512-4145-8583-cc1aa74307d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.031772] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Preparing fetch location {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 899.032037] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Fetch image to [datastore1] OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464/OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464.vmdk {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 899.032161] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Downloading stream optimized image 37eeba24-1178-4ae0-aa71-cb4902ea42c8 to [datastore1] OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464/OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464.vmdk on the data store datastore1 as vApp {{(pid=61911) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 899.032349] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Downloading image file data 37eeba24-1178-4ae0-aa71-cb4902ea42c8 to the ESX as VM named 'OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464' {{(pid=61911) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 899.095944] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251268, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100467} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.095944] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.096774] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83019921-8895-4b9c-9231-15cf3b3e80f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.124503] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.126750] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2339633-3cee-41e7-8458-0bafc4fbdd3e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.144261] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 899.144261] env[61911]: value = "resgroup-9" [ 899.144261] env[61911]: _type = "ResourcePool" [ 899.144261] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 899.144822] env[61911]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-04675a0f-e0b7-4f37-8d9a-e71ce6c3bfcc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.172681] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lease: (returnval){ [ 899.172681] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521e569b-11df-41a8-2b72-97b2c74e91be" [ 899.172681] env[61911]: _type = "HttpNfcLease" [ 899.172681] env[61911]: } obtained for vApp import into resource pool (val){ [ 899.172681] env[61911]: value = "resgroup-9" [ 899.172681] env[61911]: _type = "ResourcePool" [ 899.172681] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 899.173920] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the lease: (returnval){ [ 899.173920] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521e569b-11df-41a8-2b72-97b2c74e91be" [ 899.173920] env[61911]: _type = "HttpNfcLease" [ 899.173920] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 899.173920] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 899.173920] env[61911]: value = "task-1251269" [ 899.173920] env[61911]: _type = "Task" [ 899.173920] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.177498] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.191288] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251269, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.191288] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.191288] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521e569b-11df-41a8-2b72-97b2c74e91be" [ 899.191288] env[61911]: _type = "HttpNfcLease" [ 899.191288] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 899.269444] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.975s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.270876] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.289s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.294343] env[61911]: INFO nova.scheduler.client.report [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Deleted allocations for instance 1058f8d2-0979-4423-802d-52d91d5a2ff7 [ 899.436280] env[61911]: DEBUG nova.network.neutron [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Updated VIF entry in instance network info cache for port 3a287a5b-935a-49ef-b44f-2e53162f4440. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.436676] env[61911]: DEBUG nova.network.neutron [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Updating instance_info_cache with network_info: [{"id": "3a287a5b-935a-49ef-b44f-2e53162f4440", "address": "fa:16:3e:2d:d2:36", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a287a5b-93", "ovs_interfaceid": "3a287a5b-935a-49ef-b44f-2e53162f4440", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.470690] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "50a522ab-651a-4bb3-93d3-8866169e95e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.471026] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.471278] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "50a522ab-651a-4bb3-93d3-8866169e95e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.471557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.471656] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.475940] env[61911]: INFO nova.compute.manager [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Terminating instance [ 899.538403] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 899.538771] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0fa1dd29-61bc-44a1-9d16-742816c89722 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.547179] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 899.547179] env[61911]: value = "task-1251271" [ 899.547179] env[61911]: _type = "Task" [ 899.547179] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.557246] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251271, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.627849] env[61911]: DEBUG nova.network.neutron [-] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.680059] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.694509] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.694509] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521e569b-11df-41a8-2b72-97b2c74e91be" [ 899.694509] env[61911]: _type = "HttpNfcLease" [ 899.694509] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 899.694849] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.777260] env[61911]: INFO nova.compute.claims [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.803750] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c732f45b-442f-4c64-afb8-66d9299321db tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "1058f8d2-0979-4423-802d-52d91d5a2ff7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.049s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.943973] env[61911]: DEBUG oslo_concurrency.lockutils [req-9892baad-6e2f-4afa-b37c-e3b8e69f47fb req-915e3bd0-caef-4611-9503-d9c842249d18 service nova] Releasing lock "refresh_cache-b8cd7073-c712-4470-8ece-d144e8ce275f" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.983020] env[61911]: DEBUG nova.compute.manager [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 899.983020] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.983020] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80baf8e3-8af0-4396-a1b1-5e423e86648d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.993276] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.993276] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b37d440e-2bb1-4611-9e6c-296833fc609c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.001806] env[61911]: DEBUG oslo_vmware.api [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 900.001806] env[61911]: value = "task-1251272" [ 900.001806] env[61911]: _type = "Task" [ 900.001806] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.014009] env[61911]: DEBUG oslo_vmware.api [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.060538] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251271, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.131564] env[61911]: INFO nova.compute.manager [-] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Took 1.81 seconds to deallocate network for instance. [ 900.198040] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.207137] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251269, 'name': ReconfigVM_Task, 'duration_secs': 0.73653} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.207137] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 900.207137] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521e569b-11df-41a8-2b72-97b2c74e91be" [ 900.207137] env[61911]: _type = "HttpNfcLease" [ 900.207137] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 900.207443] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 66472f43-537d-4eb3-8d49-d40627a8809d/66472f43-537d-4eb3-8d49-d40627a8809d.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.207955] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 900.207955] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521e569b-11df-41a8-2b72-97b2c74e91be" [ 900.207955] env[61911]: _type = "HttpNfcLease" [ 900.207955] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 900.208271] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69d12d90-3e70-433d-87c3-2875f1fbd445 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.210791] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f707651-2cfb-418d-b131-d6a6c808388b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.220320] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52213ae7-0135-01f0-ddf1-e361827b7613/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 900.220515] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52213ae7-0135-01f0-ddf1-e361827b7613/disk-0.vmdk. {{(pid=61911) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 900.223259] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 900.223259] env[61911]: value = "task-1251273" [ 900.223259] env[61911]: _type = "Task" [ 900.223259] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.284384] env[61911]: INFO nova.compute.resource_tracker [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating resource usage from migration 637723a6-e295-453e-8eae-95ce6f45696e [ 900.292371] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251273, 'name': Rename_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.295726] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2d439bf0-7072-4791-8a74-2175248d9d7c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.468800] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "83616524-2530-4398-9019-148e0658720d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.469151] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "83616524-2530-4398-9019-148e0658720d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.469468] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "83616524-2530-4398-9019-148e0658720d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.469715] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "83616524-2530-4398-9019-148e0658720d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.469912] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "83616524-2530-4398-9019-148e0658720d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.472195] env[61911]: INFO nova.compute.manager [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Terminating instance [ 900.515975] env[61911]: DEBUG oslo_vmware.api [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251272, 'name': PowerOffVM_Task, 'duration_secs': 0.229732} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.517943] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.518279] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.518468] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0114d295-44c0-4ea3-98aa-72c1c2e25e46 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.554135] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38ba8b1-70f0-4835-a0fc-ff797552b0b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.564396] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251271, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.568267] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d94a8b-f442-41a4-a79b-f8670b6e89db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.609343] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8860264c-4e9e-4bfa-9fbc-ed9048002d1d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.612243] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.612455] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.612654] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleting the datastore file [datastore1] 50a522ab-651a-4bb3-93d3-8866169e95e4 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.612927] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7b6bd92-adf9-4a9a-a889-05f377312aea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.622947] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e838af16-4b0b-42d2-9330-b24fa36aeaec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.627561] env[61911]: DEBUG oslo_vmware.api [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 900.627561] env[61911]: value = "task-1251275" [ 900.627561] env[61911]: _type = "Task" [ 900.627561] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.642122] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.642637] env[61911]: DEBUG nova.compute.provider_tree [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.653551] env[61911]: DEBUG oslo_vmware.api [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.681859] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.735909] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251273, 'name': Rename_Task, 'duration_secs': 0.162702} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.736292] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.736587] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d95c144-b116-4476-8e79-d9a49741fda4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.747442] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 900.747442] env[61911]: value = "task-1251276" [ 900.747442] env[61911]: _type = "Task" [ 900.747442] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.760402] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251276, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.764930] env[61911]: DEBUG nova.compute.manager [req-1f7c91b1-b5c8-4048-b30b-e9a912910951 req-94e6dbac-f39e-4e69-9eaf-e55c08ebea64 service nova] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Received event network-vif-deleted-6531946d-4a56-42a8-9fcc-9ba3c928ad5d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 900.978617] env[61911]: DEBUG nova.compute.manager [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 900.978945] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.980852] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfc7753-408d-4557-ace3-bffd69386d09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.992766] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.994274] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d5c9341-a216-4fa9-8ef2-d367ae340803 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.003011] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 901.003011] env[61911]: value = "task-1251277" [ 901.003011] env[61911]: _type = "Task" [ 901.003011] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.017232] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.057991] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251271, 'name': CloneVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.139674] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Completed reading data from the image iterator. {{(pid=61911) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 901.139999] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52213ae7-0135-01f0-ddf1-e361827b7613/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 901.141060] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d14772a-7ef5-47fb-8f3a-74bfc770cc00 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.148450] env[61911]: DEBUG nova.scheduler.client.report [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 901.152141] env[61911]: DEBUG oslo_vmware.api [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251275, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163755} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.153016] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.153273] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.154499] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.154499] env[61911]: INFO nova.compute.manager [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Took 1.17 seconds to destroy the instance on the hypervisor. [ 901.154499] env[61911]: DEBUG oslo.service.loopingcall [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.154499] env[61911]: DEBUG nova.compute.manager [-] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 901.154499] env[61911]: DEBUG nova.network.neutron [-] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 901.157707] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52213ae7-0135-01f0-ddf1-e361827b7613/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 901.157865] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52213ae7-0135-01f0-ddf1-e361827b7613/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 901.158447] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9d4bc2b2-e50f-4627-9716-4a6827b6fc7c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.181512] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.262710] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251276, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.365182] env[61911]: DEBUG oslo_vmware.rw_handles [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52213ae7-0135-01f0-ddf1-e361827b7613/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 901.365182] env[61911]: INFO nova.virt.vmwareapi.images [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Downloaded image file data 37eeba24-1178-4ae0-aa71-cb4902ea42c8 [ 901.365182] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c958ec41-0f01-49b6-be19-008f49bdf8e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.384138] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e82b43b-3365-4f6f-9baa-ef4863cbec33 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.417156] env[61911]: INFO nova.virt.vmwareapi.images [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] The imported VM was unregistered [ 901.421024] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Caching image {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 901.421024] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating directory with path [datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.421219] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c16b893-e57a-41ac-a025-9513713d0f78 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.433563] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Created directory with path [datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.433786] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464/OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464.vmdk to [datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8/37eeba24-1178-4ae0-aa71-cb4902ea42c8.vmdk. {{(pid=61911) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 901.434101] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ec641bf1-728f-460b-bee2-95a94d340d0b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.443138] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 901.443138] env[61911]: value = "task-1251279" [ 901.443138] env[61911]: _type = "Task" [ 901.443138] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.452085] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251279, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.514486] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251277, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.538097] env[61911]: DEBUG nova.compute.manager [req-bae0247d-58e1-43fb-a775-7451295f6915 req-d1e35332-1dec-497c-8b16-b3985f3bff4b service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Received event network-vif-deleted-76870f3b-7732-4724-8dc0-38e3012a2fee {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 901.538408] env[61911]: INFO nova.compute.manager [req-bae0247d-58e1-43fb-a775-7451295f6915 req-d1e35332-1dec-497c-8b16-b3985f3bff4b service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Neutron deleted interface 76870f3b-7732-4724-8dc0-38e3012a2fee; detaching it from the instance and deleting it from the info cache [ 901.538648] env[61911]: DEBUG nova.network.neutron [req-bae0247d-58e1-43fb-a775-7451295f6915 req-d1e35332-1dec-497c-8b16-b3985f3bff4b service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.560392] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251271, 'name': CloneVM_Task, 'duration_secs': 1.612869} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.560719] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Created linked-clone VM from snapshot [ 901.561527] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb55e615-797c-491c-81fa-4a9726f10324 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.571175] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Uploading image d8718bf5-976e-48a0-b72e-8fea0db5e230 {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 901.595683] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 901.595683] env[61911]: value = "vm-269623" [ 901.595683] env[61911]: _type = "VirtualMachine" [ 901.595683] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 901.596015] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-820eef6c-53fb-4d02-96f6-7747a66c65f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.604927] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lease: (returnval){ [ 901.604927] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52563d88-87c8-5ae1-4762-073485c88124" [ 901.604927] env[61911]: _type = "HttpNfcLease" [ 901.604927] env[61911]: } obtained for exporting VM: (result){ [ 901.604927] env[61911]: value = "vm-269623" [ 901.604927] env[61911]: _type = "VirtualMachine" [ 901.604927] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 901.605336] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the lease: (returnval){ [ 901.605336] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52563d88-87c8-5ae1-4762-073485c88124" [ 901.605336] env[61911]: _type = "HttpNfcLease" [ 901.605336] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 901.615090] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.615090] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52563d88-87c8-5ae1-4762-073485c88124" [ 901.615090] env[61911]: _type = "HttpNfcLease" [ 901.615090] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 901.654090] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.383s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.654506] env[61911]: INFO nova.compute.manager [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Migrating [ 901.654797] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.654966] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.656729] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.015s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.657026] env[61911]: DEBUG nova.objects.instance [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lazy-loading 'resources' on Instance uuid 28159e75-9fe9-44c7-b5c9-534902cecbef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.686037] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.762067] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251276, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.954565] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251279, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.015873] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251277, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.018273] env[61911]: DEBUG nova.network.neutron [-] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.041517] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a6c1b8c-86e8-4aaa-94e9-8eba0773d8c8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.052725] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf5da5f-4d35-4800-a170-25f1cb0bec33 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.098849] env[61911]: DEBUG nova.compute.manager [req-bae0247d-58e1-43fb-a775-7451295f6915 req-d1e35332-1dec-497c-8b16-b3985f3bff4b service nova] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Detach interface failed, port_id=76870f3b-7732-4724-8dc0-38e3012a2fee, reason: Instance 50a522ab-651a-4bb3-93d3-8866169e95e4 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 902.114174] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.114174] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52563d88-87c8-5ae1-4762-073485c88124" [ 902.114174] env[61911]: _type = "HttpNfcLease" [ 902.114174] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.114541] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.114541] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52563d88-87c8-5ae1-4762-073485c88124" [ 902.114541] env[61911]: _type = "HttpNfcLease" [ 902.114541] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.115343] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c79642-6a0f-4c39-9d25-7b240ac5d826 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.125023] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204744f-d1fe-48ad-2233-29477f556e58/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.125328] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204744f-d1fe-48ad-2233-29477f556e58/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.187103] env[61911]: INFO nova.compute.rpcapi [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 902.187103] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.220618] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.258245] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ef4aa067-1c95-42e2-b134-3daab10dc542 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.269265] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251276, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.459134] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251279, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.493331] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3738ae15-e856-49c1-bb19-b264c8e0bbdf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.504433] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca197261-f6ce-4609-acc2-32c98664983e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.540512] env[61911]: INFO nova.compute.manager [-] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Took 1.39 seconds to deallocate network for instance. [ 902.546221] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995a34b9-a7f2-4a05-b2c6-b348c62db7db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.549231] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251277, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.561470] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60663bb8-3410-40fc-8d45-122ab181ae6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.578103] env[61911]: DEBUG nova.compute.provider_tree [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.711262] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.711888] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.712076] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.712318] env[61911]: DEBUG nova.network.neutron [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.764594] env[61911]: DEBUG oslo_vmware.api [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251276, 'name': PowerOnVM_Task, 'duration_secs': 1.7478} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.765040] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.765342] env[61911]: DEBUG nova.compute.manager [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 902.766329] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dcac50-a7a4-40ae-8dc9-07377fe9b61a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.956185] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251279, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.020390] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251277, 'name': PowerOffVM_Task, 'duration_secs': 1.820743} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.020594] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.020905] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.021416] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b070b6e-3c5c-42a2-8436-7e58884c0a59 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.055091] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.082126] env[61911]: DEBUG nova.scheduler.client.report [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.157404] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.157813] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.158088] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Deleting the datastore file [datastore1] 83616524-2530-4398-9019-148e0658720d {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.158490] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab8a5eb1-c819-4c77-aaec-3544414b884f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.167181] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for the task: (returnval){ [ 903.167181] env[61911]: value = "task-1251282" [ 903.167181] env[61911]: _type = "Task" [ 903.167181] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.176695] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.208549] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.291400] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.457597] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251279, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.587992] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.593093] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.537s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.593093] env[61911]: DEBUG nova.objects.instance [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lazy-loading 'resources' on Instance uuid 50a522ab-651a-4bb3-93d3-8866169e95e4 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.624602] env[61911]: INFO nova.scheduler.client.report [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Deleted allocations for instance 28159e75-9fe9-44c7-b5c9-534902cecbef [ 903.667668] env[61911]: DEBUG nova.network.neutron [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance_info_cache with network_info: [{"id": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "address": "fa:16:3e:c2:34:3f", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7ff300-bd", "ovs_interfaceid": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.680586] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.708929] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.956489] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251279, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.135138] env[61911]: DEBUG oslo_concurrency.lockutils [None req-aea5c99f-1026-4953-941e-e6f0f457e4c1 tempest-VolumesAdminNegativeTest-447687182 tempest-VolumesAdminNegativeTest-447687182-project-member] Lock "28159e75-9fe9-44c7-b5c9-534902cecbef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.475s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.175760] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.186098] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.212768] env[61911]: DEBUG oslo_vmware.api [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251267, 'name': ReconfigVM_Task, 'duration_secs': 5.99923} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.213444] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.213530] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Reconfigured VM to detach interface {{(pid=61911) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 904.339401] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31450d74-89fe-413f-b0c2-2138227b1b92 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.352884] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b63113-860a-4a72-9909-dcacb72ef4fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.388435] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39d91b3-3aeb-446e-a112-7facf7e6d9fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.398271] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277513b0-e35b-48bc-b63e-f26640683cfa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.413098] env[61911]: DEBUG nova.compute.provider_tree [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.457686] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251279, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.569282} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.457797] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464/OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464.vmdk to [datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8/37eeba24-1178-4ae0-aa71-cb4902ea42c8.vmdk. [ 904.458011] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Cleaning up location [datastore1] OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 904.458468] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_e14b7810-6264-455c-a82f-9ab0fe8b2464 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.458759] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-768be4dc-c9cd-4826-aa69-8dce3df8c15f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.468217] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 904.468217] env[61911]: value = "task-1251283" [ 904.468217] env[61911]: _type = "Task" [ 904.468217] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.476162] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.686976] env[61911]: DEBUG oslo_vmware.api [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Task: {'id': task-1251282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.182066} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.687887] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.688493] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.688863] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.689365] env[61911]: INFO nova.compute.manager [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] [instance: 83616524-2530-4398-9019-148e0658720d] Took 3.71 seconds to destroy the instance on the hypervisor. [ 904.689897] env[61911]: DEBUG oslo.service.loopingcall [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.690278] env[61911]: DEBUG nova.compute.manager [-] [instance: 83616524-2530-4398-9019-148e0658720d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 904.690521] env[61911]: DEBUG nova.network.neutron [-] [instance: 83616524-2530-4398-9019-148e0658720d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.921082] env[61911]: DEBUG nova.scheduler.client.report [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 904.984035] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.066633} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.984035] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.984035] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8/37eeba24-1178-4ae0-aa71-cb4902ea42c8.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.984035] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8/37eeba24-1178-4ae0-aa71-cb4902ea42c8.vmdk to [datastore1] b8cd7073-c712-4470-8ece-d144e8ce275f/b8cd7073-c712-4470-8ece-d144e8ce275f.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 904.984035] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-420c9594-da08-4825-903b-4607e0cc095a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.994241] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 904.994241] env[61911]: value = "task-1251284" [ 904.994241] env[61911]: _type = "Task" [ 904.994241] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.007104] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251284, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.052935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.052935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.052935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.052935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.052935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.055777] env[61911]: INFO nova.compute.manager [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Terminating instance [ 905.131558] env[61911]: DEBUG nova.compute.manager [req-e9a2ab3d-0cc3-4107-9dea-d841f6aebdfc req-53ef3a18-4950-4efa-a49f-e30745552334 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Received event network-vif-deleted-28654114-34d7-4037-8226-713299d2e033 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 905.131862] env[61911]: INFO nova.compute.manager [req-e9a2ab3d-0cc3-4107-9dea-d841f6aebdfc req-53ef3a18-4950-4efa-a49f-e30745552334 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Neutron deleted interface 28654114-34d7-4037-8226-713299d2e033; detaching it from the instance and deleting it from the info cache [ 905.132124] env[61911]: DEBUG nova.network.neutron [req-e9a2ab3d-0cc3-4107-9dea-d841f6aebdfc req-53ef3a18-4950-4efa-a49f-e30745552334 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.424394] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.428164] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.136s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.428450] env[61911]: DEBUG nova.objects.instance [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 905.457053] env[61911]: INFO nova.scheduler.client.report [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted allocations for instance 50a522ab-651a-4bb3-93d3-8866169e95e4 [ 905.514719] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251284, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.564036] env[61911]: DEBUG nova.compute.manager [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 905.564036] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 905.564036] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44004067-c970-4e7b-a70f-28db7375563d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.576811] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.579764] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0c4ae8a-83c5-4ce3-aa8c-f233458ecb01 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.589364] env[61911]: DEBUG oslo_vmware.api [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 905.589364] env[61911]: value = "task-1251285" [ 905.589364] env[61911]: _type = "Task" [ 905.589364] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.603613] env[61911]: DEBUG oslo_vmware.api [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251285, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.610024] env[61911]: DEBUG nova.network.neutron [-] [instance: 83616524-2530-4398-9019-148e0658720d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.635254] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7dce60b2-9570-4b0d-9d65-2bbbbba7f431 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.647106] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b602707a-66cc-47d0-8b41-3945905f7f03 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.684733] env[61911]: DEBUG nova.compute.manager [req-e9a2ab3d-0cc3-4107-9dea-d841f6aebdfc req-53ef3a18-4950-4efa-a49f-e30745552334 service nova] [instance: 83616524-2530-4398-9019-148e0658720d] Detach interface failed, port_id=28654114-34d7-4037-8226-713299d2e033, reason: Instance 83616524-2530-4398-9019-148e0658720d could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 905.696709] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3474c56-12c7-42f4-92c4-fd36c459ec24 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.721564] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance 'a35a11f1-3d4b-439e-9517-fa5a9f43af60' progress to 0 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 905.835999] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.836360] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquired lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.836644] env[61911]: DEBUG nova.network.neutron [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.967522] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f2e0d46f-42e0-46bf-ba96-6839b116e612 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "50a522ab-651a-4bb3-93d3-8866169e95e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.496s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.010821] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251284, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.047406] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.047753] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.047998] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.048220] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.048510] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.051708] env[61911]: INFO nova.compute.manager [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Terminating instance [ 906.101537] env[61911]: DEBUG oslo_vmware.api [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251285, 'name': PowerOffVM_Task, 'duration_secs': 0.287223} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.101868] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.102058] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.102336] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-110fa68b-d714-4c83-a330-cc753cfe5573 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.111667] env[61911]: INFO nova.compute.manager [-] [instance: 83616524-2530-4398-9019-148e0658720d] Took 1.42 seconds to deallocate network for instance. [ 906.183498] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.183753] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.183922] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleting the datastore file [datastore2] 9c2e75d8-b307-43bb-84a8-604cd5edbc39 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.186181] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4957ea0-e06a-4105-b3c4-cdbe55e0e1fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.197443] env[61911]: DEBUG oslo_vmware.api [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 906.197443] env[61911]: value = "task-1251287" [ 906.197443] env[61911]: _type = "Task" [ 906.197443] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.208485] env[61911]: DEBUG oslo_vmware.api [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251287, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.228549] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.228931] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b62a6ed-8aa5-4d81-882f-71af3d9c9a5f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.239432] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 906.239432] env[61911]: value = "task-1251288" [ 906.239432] env[61911]: _type = "Task" [ 906.239432] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.251405] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.441644] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3c5ee072-3698-43c9-8ef0-f0432e4d4bae tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.511715] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251284, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.557566] env[61911]: DEBUG nova.compute.manager [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 906.557896] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.558890] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9e7024-cd8d-431e-a1dd-939aebcd5433 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.568459] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.568787] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-144538df-fb78-49ad-9d15-ada6f4601759 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.578497] env[61911]: DEBUG oslo_vmware.api [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 906.578497] env[61911]: value = "task-1251289" [ 906.578497] env[61911]: _type = "Task" [ 906.578497] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.591926] env[61911]: DEBUG oslo_vmware.api [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251289, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.619516] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.620039] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.620357] env[61911]: DEBUG nova.objects.instance [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lazy-loading 'resources' on Instance uuid 83616524-2530-4398-9019-148e0658720d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.622835] env[61911]: INFO nova.network.neutron [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Port c34b4a10-4268-487c-bb19-1ef7beda1f06 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 906.623278] env[61911]: DEBUG nova.network.neutron [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [{"id": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "address": "fa:16:3e:c0:05:0f", "network": {"id": "87858d49-056a-4528-87e1-d07b16c7f333", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-292525191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f40f512384d4fedb75bfc53f3195243", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde067cc9-82", "ovs_interfaceid": "de067cc9-827d-4e43-a53e-6eda53bf9d1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.708899] env[61911]: DEBUG oslo_vmware.api [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251287, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.331552} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.710811] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.710811] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 906.710811] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 906.710811] env[61911]: INFO nova.compute.manager [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Took 1.15 seconds to destroy the instance on the hypervisor. [ 906.710811] env[61911]: DEBUG oslo.service.loopingcall [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.711200] env[61911]: DEBUG nova.compute.manager [-] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 906.711200] env[61911]: DEBUG nova.network.neutron [-] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 906.750967] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251288, 'name': PowerOffVM_Task, 'duration_secs': 0.283802} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.751439] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.751714] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance 'a35a11f1-3d4b-439e-9517-fa5a9f43af60' progress to 17 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 907.010383] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251284, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.095557] env[61911]: DEBUG oslo_vmware.api [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251289, 'name': PowerOffVM_Task, 'duration_secs': 0.271581} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.095557] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.095557] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 907.095557] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-215b34e9-7e28-4a56-8afa-8272fb94b267 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.130301] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Releasing lock "refresh_cache-e3ad3416-423c-4687-aa35-2fb2ef67ec64" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.170782] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 907.171118] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 907.171330] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleting the datastore file [datastore2] e3ad3416-423c-4687-aa35-2fb2ef67ec64 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.171659] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d262a11-592a-4fa5-9ad2-2f113d752293 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.183917] env[61911]: DEBUG oslo_vmware.api [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 907.183917] env[61911]: value = "task-1251291" [ 907.183917] env[61911]: _type = "Task" [ 907.183917] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.194354] env[61911]: DEBUG oslo_vmware.api [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.247212] env[61911]: DEBUG nova.compute.manager [req-76948293-71a2-4d63-a54a-d951dba0fa0a req-bb420d9c-1549-4b59-bde3-c91ed8b5c2a0 service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Received event network-vif-deleted-54a15f3e-c3d1-45b1-ac13-19e74bed26c6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 907.247212] env[61911]: INFO nova.compute.manager [req-76948293-71a2-4d63-a54a-d951dba0fa0a req-bb420d9c-1549-4b59-bde3-c91ed8b5c2a0 service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Neutron deleted interface 54a15f3e-c3d1-45b1-ac13-19e74bed26c6; detaching it from the instance and deleting it from the info cache [ 907.247212] env[61911]: DEBUG nova.network.neutron [req-76948293-71a2-4d63-a54a-d951dba0fa0a req-bb420d9c-1549-4b59-bde3-c91ed8b5c2a0 service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.267313] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:34Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 907.267639] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.267810] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 907.268014] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.268199] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 907.268357] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 907.268667] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 907.268846] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 907.269035] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 907.269216] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 907.269398] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 907.278571] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49acaa77-7f0c-4d37-8ca7-283e670fda73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.297970] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 907.297970] env[61911]: value = "task-1251292" [ 907.297970] env[61911]: _type = "Task" [ 907.297970] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.308483] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251292, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.417030] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "75577b68-b012-43d5-abdb-bd5f54c249d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.418564] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.486580] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f61e9d6-7dee-4bb4-b8c7-34c13ced7d31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.495066] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5992d03e-208e-49f9-9714-6443efad0ce1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.506898] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251284, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.412565} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.530014] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37eeba24-1178-4ae0-aa71-cb4902ea42c8/37eeba24-1178-4ae0-aa71-cb4902ea42c8.vmdk to [datastore1] b8cd7073-c712-4470-8ece-d144e8ce275f/b8cd7073-c712-4470-8ece-d144e8ce275f.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 907.531203] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edd3a2a-552b-4438-a0a6-8586d216207d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.534218] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81dd8f3-9146-47f3-ac4d-887b18e7bebd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.553910] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402e39b6-b05f-4169-9e72-51677b7edaee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.568585] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] b8cd7073-c712-4470-8ece-d144e8ce275f/b8cd7073-c712-4470-8ece-d144e8ce275f.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.568585] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a331efc-9b05-43ad-837d-fa0201e87b2c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.583194] env[61911]: DEBUG nova.network.neutron [-] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.592717] env[61911]: DEBUG nova.compute.provider_tree [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.595820] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 907.595820] env[61911]: value = "task-1251293" [ 907.595820] env[61911]: _type = "Task" [ 907.595820] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.605422] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251293, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.639035] env[61911]: DEBUG oslo_concurrency.lockutils [None req-daa5ae7f-b354-4ffe-a138-4909adcf8ec7 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "interface-e3ad3416-423c-4687-aa35-2fb2ef67ec64-c34b4a10-4268-487c-bb19-1ef7beda1f06" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.069s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.695401] env[61911]: DEBUG oslo_vmware.api [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.450998} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.696049] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.696049] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.696178] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.696285] env[61911]: INFO nova.compute.manager [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Took 1.14 seconds to destroy the instance on the hypervisor. [ 907.696665] env[61911]: DEBUG oslo.service.loopingcall [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.696864] env[61911]: DEBUG nova.compute.manager [-] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 907.696964] env[61911]: DEBUG nova.network.neutron [-] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.749927] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9de3f7db-bee3-42e0-93e3-0f5dea599010 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.759315] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89c221c-b247-41bf-9e2c-d2e203500a22 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.793142] env[61911]: DEBUG nova.compute.manager [req-76948293-71a2-4d63-a54a-d951dba0fa0a req-bb420d9c-1549-4b59-bde3-c91ed8b5c2a0 service nova] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Detach interface failed, port_id=54a15f3e-c3d1-45b1-ac13-19e74bed26c6, reason: Instance 9c2e75d8-b307-43bb-84a8-604cd5edbc39 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 907.808672] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251292, 'name': ReconfigVM_Task, 'duration_secs': 0.318336} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.809016] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance 'a35a11f1-3d4b-439e-9517-fa5a9f43af60' progress to 33 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 907.920895] env[61911]: DEBUG nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 908.095992] env[61911]: INFO nova.compute.manager [-] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Took 1.38 seconds to deallocate network for instance. [ 908.097484] env[61911]: DEBUG nova.scheduler.client.report [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 908.128613] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251293, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.317492] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 908.317916] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.318234] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 908.318585] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.318848] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 908.319104] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 908.319415] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 908.319666] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 908.319920] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 908.320184] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 908.320453] env[61911]: DEBUG nova.virt.hardware [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 908.328945] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Reconfiguring VM instance instance-0000004e to detach disk 2000 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 908.329375] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68151b3d-ce31-4f66-94b4-37014b98f401 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.350586] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 908.350586] env[61911]: value = "task-1251294" [ 908.350586] env[61911]: _type = "Task" [ 908.350586] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.359786] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251294, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.450429] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.607349] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.610073] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.160s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.612441] env[61911]: INFO nova.compute.claims [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.625327] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.625816] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251293, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.633907] env[61911]: INFO nova.scheduler.client.report [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Deleted allocations for instance 83616524-2530-4398-9019-148e0658720d [ 908.863106] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251294, 'name': ReconfigVM_Task, 'duration_secs': 0.336142} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.865064] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Reconfigured VM instance instance-0000004e to detach disk 2000 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 908.865064] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7570684-344d-4353-9634-206a937c437d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.895990] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] a35a11f1-3d4b-439e-9517-fa5a9f43af60/a35a11f1-3d4b-439e-9517-fa5a9f43af60.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.896383] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4e74577-d70e-4074-99f6-8273903bdda1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.915642] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 908.915642] env[61911]: value = "task-1251295" [ 908.915642] env[61911]: _type = "Task" [ 908.915642] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.924680] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.986655] env[61911]: DEBUG nova.network.neutron [-] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.114762] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251293, 'name': ReconfigVM_Task, 'duration_secs': 1.11181} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.115225] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Reconfigured VM instance instance-00000050 to attach disk [datastore1] b8cd7073-c712-4470-8ece-d144e8ce275f/b8cd7073-c712-4470-8ece-d144e8ce275f.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.115686] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fee361c-1603-4bfc-be19-15b03ca7bf25 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.129915] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 909.129915] env[61911]: value = "task-1251296" [ 909.129915] env[61911]: _type = "Task" [ 909.129915] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.142076] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251296, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.142575] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8d9aa6b7-ca5e-4684-adf6-4eb99bc65ff0 tempest-ServerRescueTestJSON-633049658 tempest-ServerRescueTestJSON-633049658-project-member] Lock "83616524-2530-4398-9019-148e0658720d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.673s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.292082] env[61911]: DEBUG nova.compute.manager [req-13be1ba9-2ae2-4418-88ce-1b0d4daa79a6 req-5726a708-95f3-491e-9eef-74a225de1469 service nova] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Received event network-vif-deleted-de067cc9-827d-4e43-a53e-6eda53bf9d1e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 909.430565] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251295, 'name': ReconfigVM_Task, 'duration_secs': 0.427329} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.431042] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Reconfigured VM instance instance-0000004e to attach disk [datastore1] a35a11f1-3d4b-439e-9517-fa5a9f43af60/a35a11f1-3d4b-439e-9517-fa5a9f43af60.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.431390] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance 'a35a11f1-3d4b-439e-9517-fa5a9f43af60' progress to 50 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 909.492124] env[61911]: INFO nova.compute.manager [-] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Took 1.79 seconds to deallocate network for instance. [ 909.639758] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251296, 'name': Rename_Task, 'duration_secs': 0.168993} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.640084] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 909.640364] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-365fdc60-1986-43d3-8bad-7f85fa4e135a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.650551] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 909.650551] env[61911]: value = "task-1251297" [ 909.650551] env[61911]: _type = "Task" [ 909.650551] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.663918] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.892307] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7914e1-30b9-498f-9002-d1ac2917cf7c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.901585] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f11a894-22e0-4ebb-8619-5c2051283d45 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.933292] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d067f33-a392-47b9-becb-fcea5bb9b92a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.942737] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97c78bc-52a8-41d1-8766-35f60e419a62 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.946701] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc96f9b-0033-4efe-96dc-68818156a584 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.962911] env[61911]: DEBUG nova.compute.provider_tree [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.979023] env[61911]: DEBUG nova.scheduler.client.report [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 909.983786] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dce3e6e-5b1a-4630-a77f-3713c53bacc0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.009082] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.009415] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance 'a35a11f1-3d4b-439e-9517-fa5a9f43af60' progress to 67 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 910.163105] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251297, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.489887] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.880s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.492949] env[61911]: DEBUG nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 910.495042] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.870s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.498375] env[61911]: DEBUG nova.objects.instance [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lazy-loading 'resources' on Instance uuid 9c2e75d8-b307-43bb-84a8-604cd5edbc39 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.553504] env[61911]: DEBUG nova.network.neutron [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Port 4e7ff300-bdd2-43d6-9fe5-646ffb508e5f binding to destination host cpu-1 is already ACTIVE {{(pid=61911) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 910.662632] env[61911]: DEBUG oslo_vmware.api [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251297, 'name': PowerOnVM_Task, 'duration_secs': 0.75244} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.662996] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 910.663336] env[61911]: INFO nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Took 15.52 seconds to spawn the instance on the hypervisor. [ 910.663669] env[61911]: DEBUG nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 910.664524] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d015e12-6465-47e4-8a0b-6c254df66ecd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.001917] env[61911]: DEBUG nova.compute.utils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.008864] env[61911]: DEBUG nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 911.009289] env[61911]: DEBUG nova.network.neutron [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.071371] env[61911]: DEBUG nova.policy [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe50c877fcef489cb798fafe73caaf8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813b3181120245e594809d096a621675', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.188354] env[61911]: INFO nova.compute.manager [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Took 29.63 seconds to build instance. [ 911.297062] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d696eb82-7efa-4f04-a3ed-f9b477d63c99 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.305933] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2088eb63-db5f-47fd-b7ba-e931e4287334 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.343262] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913eef57-a7e0-41b8-ae24-9cbb85136c23 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.355048] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade74e2d-efa6-4061-b23d-46b07a9022e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.367519] env[61911]: DEBUG nova.compute.provider_tree [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.508993] env[61911]: DEBUG nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 911.586915] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.587251] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.587476] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.601988] env[61911]: DEBUG nova.network.neutron [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Successfully created port: 4da57eb9-f431-463c-b1fc-292ec04a8cac {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.691430] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f3a3258a-b09d-4429-81d3-84a7ba09125a tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.143s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.870974] env[61911]: DEBUG nova.scheduler.client.report [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 912.376744] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.882s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.380540] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.371s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.381650] env[61911]: DEBUG nova.objects.instance [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lazy-loading 'resources' on Instance uuid e3ad3416-423c-4687-aa35-2fb2ef67ec64 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.419583] env[61911]: INFO nova.scheduler.client.report [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted allocations for instance 9c2e75d8-b307-43bb-84a8-604cd5edbc39 [ 912.519027] env[61911]: DEBUG nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 912.546830] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 912.547136] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.547319] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 912.547508] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.547665] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 912.547964] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 912.548720] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 912.548720] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 912.548929] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 912.549125] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 912.549309] env[61911]: DEBUG nova.virt.hardware [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 912.550478] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7c19d8-4823-4c96-89d7-ff76a972d19e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.559703] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38a5395-e99a-4558-9ffb-d12d5313f79c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.688431] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.688718] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.689016] env[61911]: DEBUG nova.network.neutron [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.741432] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204744f-d1fe-48ad-2233-29477f556e58/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 912.742464] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf90957-930d-4200-a704-b78c965d1886 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.749059] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204744f-d1fe-48ad-2233-29477f556e58/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 912.749246] env[61911]: ERROR oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204744f-d1fe-48ad-2233-29477f556e58/disk-0.vmdk due to incomplete transfer. [ 912.749470] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b8cc5b3d-88bc-4be6-9bf0-6f889a2cc10f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.756516] env[61911]: DEBUG oslo_vmware.rw_handles [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204744f-d1fe-48ad-2233-29477f556e58/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 912.757057] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Uploaded image d8718bf5-976e-48a0-b72e-8fea0db5e230 to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 912.759636] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 912.759636] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c0c7cb33-0dee-444e-b15b-c171e35e2c8f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.774992] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 912.774992] env[61911]: value = "task-1251298" [ 912.774992] env[61911]: _type = "Task" [ 912.774992] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.783014] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251298, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.935433] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e9088a4c-89f8-477a-b5b5-7ed904dfa29a tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "9c2e75d8-b307-43bb-84a8-604cd5edbc39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.884s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.017186] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "b8cd7073-c712-4470-8ece-d144e8ce275f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.017186] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.017337] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "b8cd7073-c712-4470-8ece-d144e8ce275f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.017520] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.018037] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.020146] env[61911]: INFO nova.compute.manager [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Terminating instance [ 913.049847] env[61911]: DEBUG nova.compute.manager [req-f33e733c-f4a2-414c-8b16-33146a7addfe req-7edaa285-4ee9-4642-9b33-2c7bd5f9691d service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Received event network-vif-plugged-4da57eb9-f431-463c-b1fc-292ec04a8cac {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 913.050300] env[61911]: DEBUG oslo_concurrency.lockutils [req-f33e733c-f4a2-414c-8b16-33146a7addfe req-7edaa285-4ee9-4642-9b33-2c7bd5f9691d service nova] Acquiring lock "75577b68-b012-43d5-abdb-bd5f54c249d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.050466] env[61911]: DEBUG oslo_concurrency.lockutils [req-f33e733c-f4a2-414c-8b16-33146a7addfe req-7edaa285-4ee9-4642-9b33-2c7bd5f9691d service nova] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.050644] env[61911]: DEBUG oslo_concurrency.lockutils [req-f33e733c-f4a2-414c-8b16-33146a7addfe req-7edaa285-4ee9-4642-9b33-2c7bd5f9691d service nova] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.050830] env[61911]: DEBUG nova.compute.manager [req-f33e733c-f4a2-414c-8b16-33146a7addfe req-7edaa285-4ee9-4642-9b33-2c7bd5f9691d service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] No waiting events found dispatching network-vif-plugged-4da57eb9-f431-463c-b1fc-292ec04a8cac {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 913.051021] env[61911]: WARNING nova.compute.manager [req-f33e733c-f4a2-414c-8b16-33146a7addfe req-7edaa285-4ee9-4642-9b33-2c7bd5f9691d service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Received unexpected event network-vif-plugged-4da57eb9-f431-463c-b1fc-292ec04a8cac for instance with vm_state building and task_state spawning. [ 913.164744] env[61911]: DEBUG nova.network.neutron [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Successfully updated port: 4da57eb9-f431-463c-b1fc-292ec04a8cac {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.222340] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a9ca3a-b499-4aef-8572-cf07628f7feb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.234547] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f8459f-a1ff-4e43-8ff1-7811ef44e5b5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.266887] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e459e64-0090-4726-8cb5-0fec343101ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.275093] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37948a3e-04e3-47f3-97ad-e21a68b6ab98 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.294925] env[61911]: DEBUG nova.compute.provider_tree [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.299308] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251298, 'name': Destroy_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.446649] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "613544bc-92b9-4c8f-8a7b-f02205f347ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.448197] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.448197] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "613544bc-92b9-4c8f-8a7b-f02205f347ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.448477] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.448899] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.451794] env[61911]: DEBUG nova.network.neutron [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance_info_cache with network_info: [{"id": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "address": "fa:16:3e:c2:34:3f", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7ff300-bd", "ovs_interfaceid": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.456019] env[61911]: INFO nova.compute.manager [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Terminating instance [ 913.526381] env[61911]: DEBUG nova.compute.manager [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 913.529018] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.529018] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de78bcf7-8ccb-482b-a44f-ab3ccb92b2f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.542082] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.542082] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1001e0c-cf01-4d7f-91c3-ceba437196e0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.547859] env[61911]: DEBUG oslo_vmware.api [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 913.547859] env[61911]: value = "task-1251299" [ 913.547859] env[61911]: _type = "Task" [ 913.547859] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.556021] env[61911]: DEBUG oslo_vmware.api [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.673169] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "refresh_cache-75577b68-b012-43d5-abdb-bd5f54c249d5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.673334] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "refresh_cache-75577b68-b012-43d5-abdb-bd5f54c249d5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.673478] env[61911]: DEBUG nova.network.neutron [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.791661] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251298, 'name': Destroy_Task, 'duration_secs': 0.566527} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.791843] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Destroyed the VM [ 913.792192] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 913.792453] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3e18d988-a248-4b36-8af8-7d7c34e84b5a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.799052] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 913.799052] env[61911]: value = "task-1251300" [ 913.799052] env[61911]: _type = "Task" [ 913.799052] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.803217] env[61911]: DEBUG nova.scheduler.client.report [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 913.810617] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251300, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.957594] env[61911]: DEBUG oslo_concurrency.lockutils [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.964655] env[61911]: DEBUG nova.compute.manager [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 913.964655] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.964655] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7499a252-b4a5-42e6-b10f-957822ca906b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.971083] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.971345] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-677ba318-acea-4f2a-8fa1-a8634ef14bac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.977198] env[61911]: DEBUG oslo_vmware.api [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 913.977198] env[61911]: value = "task-1251301" [ 913.977198] env[61911]: _type = "Task" [ 913.977198] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.985178] env[61911]: DEBUG oslo_vmware.api [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.060677] env[61911]: DEBUG oslo_vmware.api [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.205299] env[61911]: DEBUG nova.network.neutron [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.310811] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.313058] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251300, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.334351] env[61911]: INFO nova.scheduler.client.report [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleted allocations for instance e3ad3416-423c-4687-aa35-2fb2ef67ec64 [ 914.417577] env[61911]: DEBUG nova.network.neutron [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Updating instance_info_cache with network_info: [{"id": "4da57eb9-f431-463c-b1fc-292ec04a8cac", "address": "fa:16:3e:20:82:03", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4da57eb9-f4", "ovs_interfaceid": "4da57eb9-f431-463c-b1fc-292ec04a8cac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.489386] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.489754] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.492768] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd23d3f-4091-432b-a1a3-18eeae149a43 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.499284] env[61911]: DEBUG oslo_vmware.api [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251301, 'name': PowerOffVM_Task, 'duration_secs': 0.184782} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.499999] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.500571] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.501550] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff459c54-ef35-4b32-869e-f439c2a02cbc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.520707] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36255acd-d770-46ee-b32c-bdc4e3252763 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.528922] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance 'a35a11f1-3d4b-439e-9517-fa5a9f43af60' progress to 83 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 914.560229] env[61911]: DEBUG oslo_vmware.api [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251299, 'name': PowerOffVM_Task, 'duration_secs': 1.000386} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.560533] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.560709] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.561020] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4ea0843-1007-4e81-9d0d-9ae36e96a6a5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.593465] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.593726] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.593948] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleting the datastore file [datastore1] 613544bc-92b9-4c8f-8a7b-f02205f347ff {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.594231] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-686b871c-e0a7-46af-a5e8-73c3fe4fc668 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.603021] env[61911]: DEBUG oslo_vmware.api [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 914.603021] env[61911]: value = "task-1251304" [ 914.603021] env[61911]: _type = "Task" [ 914.603021] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.618181] env[61911]: DEBUG oslo_vmware.api [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.622159] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.622421] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.622671] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleting the datastore file [datastore1] b8cd7073-c712-4470-8ece-d144e8ce275f {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.623407] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d60d3af7-de4c-4013-82db-336d1360776d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.629477] env[61911]: DEBUG oslo_vmware.api [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 914.629477] env[61911]: value = "task-1251305" [ 914.629477] env[61911]: _type = "Task" [ 914.629477] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.638451] env[61911]: DEBUG oslo_vmware.api [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251305, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.809453] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251300, 'name': RemoveSnapshot_Task, 'duration_secs': 0.964381} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.809741] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 914.810067] env[61911]: DEBUG nova.compute.manager [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.810833] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f26016f-3821-4a80-9e4f-3ba726b05c39 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.845778] env[61911]: DEBUG oslo_concurrency.lockutils [None req-75a55270-f093-4de6-8be2-c088d7eb76b5 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e3ad3416-423c-4687-aa35-2fb2ef67ec64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.798s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.924056] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "refresh_cache-75577b68-b012-43d5-abdb-bd5f54c249d5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.924300] env[61911]: DEBUG nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Instance network_info: |[{"id": "4da57eb9-f431-463c-b1fc-292ec04a8cac", "address": "fa:16:3e:20:82:03", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4da57eb9-f4", "ovs_interfaceid": "4da57eb9-f431-463c-b1fc-292ec04a8cac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 914.924790] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:82:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4da57eb9-f431-463c-b1fc-292ec04a8cac', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.933023] env[61911]: DEBUG oslo.service.loopingcall [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.933132] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.933368] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54c5b9e3-211e-4023-8367-22c741cac612 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.956126] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.956126] env[61911]: value = "task-1251306" [ 914.956126] env[61911]: _type = "Task" [ 914.956126] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.964574] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251306, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.993411] env[61911]: DEBUG nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 915.036580] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.037272] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93fd7309-3ce4-4020-a60c-3739ed4ed955 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.046136] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 915.046136] env[61911]: value = "task-1251307" [ 915.046136] env[61911]: _type = "Task" [ 915.046136] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.054787] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.114897] env[61911]: DEBUG oslo_vmware.api [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154756} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.115068] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.115356] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.115704] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.116068] env[61911]: INFO nova.compute.manager [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Took 1.15 seconds to destroy the instance on the hypervisor. [ 915.116517] env[61911]: DEBUG oslo.service.loopingcall [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.116842] env[61911]: DEBUG nova.compute.manager [-] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 915.117034] env[61911]: DEBUG nova.network.neutron [-] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 915.142536] env[61911]: DEBUG oslo_vmware.api [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251305, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146685} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.142828] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.143561] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.143842] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.144068] env[61911]: INFO nova.compute.manager [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Took 1.62 seconds to destroy the instance on the hypervisor. [ 915.144380] env[61911]: DEBUG oslo.service.loopingcall [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.144574] env[61911]: DEBUG nova.compute.manager [-] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 915.144682] env[61911]: DEBUG nova.network.neutron [-] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 915.323184] env[61911]: INFO nova.compute.manager [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Shelve offloading [ 915.353735] env[61911]: DEBUG nova.compute.manager [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Received event network-changed-4da57eb9-f431-463c-b1fc-292ec04a8cac {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 915.353735] env[61911]: DEBUG nova.compute.manager [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Refreshing instance network info cache due to event network-changed-4da57eb9-f431-463c-b1fc-292ec04a8cac. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 915.353735] env[61911]: DEBUG oslo_concurrency.lockutils [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] Acquiring lock "refresh_cache-75577b68-b012-43d5-abdb-bd5f54c249d5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.353735] env[61911]: DEBUG oslo_concurrency.lockutils [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] Acquired lock "refresh_cache-75577b68-b012-43d5-abdb-bd5f54c249d5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.355686] env[61911]: DEBUG nova.network.neutron [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Refreshing network info cache for port 4da57eb9-f431-463c-b1fc-292ec04a8cac {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.396332] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.396658] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.396877] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.397097] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.397291] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.399767] env[61911]: INFO nova.compute.manager [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Terminating instance [ 915.467017] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251306, 'name': CreateVM_Task, 'duration_secs': 0.342481} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.467455] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.468410] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.468614] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.469724] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.469953] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7945c85-dbbd-4aa2-89a0-bab142f105cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.475801] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 915.475801] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f6d5c5-00bf-a6cf-4db8-5a4fcde6dbc8" [ 915.475801] env[61911]: _type = "Task" [ 915.475801] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.484062] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f6d5c5-00bf-a6cf-4db8-5a4fcde6dbc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.523482] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.523799] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.525293] env[61911]: INFO nova.compute.claims [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.562357] env[61911]: DEBUG oslo_vmware.api [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251307, 'name': PowerOnVM_Task, 'duration_secs': 0.474518} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.563989] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.566881] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-51075921-e6b7-4316-8acb-0f41e9f59711 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance 'a35a11f1-3d4b-439e-9517-fa5a9f43af60' progress to 100 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 915.594217] env[61911]: DEBUG nova.compute.manager [req-98e973dc-38a5-472f-8799-f799c9ef3eaa req-b31f463e-78e4-400c-acc5-b7a9b4eb67bf service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Received event network-vif-deleted-3a287a5b-935a-49ef-b44f-2e53162f4440 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 915.594311] env[61911]: INFO nova.compute.manager [req-98e973dc-38a5-472f-8799-f799c9ef3eaa req-b31f463e-78e4-400c-acc5-b7a9b4eb67bf service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Neutron deleted interface 3a287a5b-935a-49ef-b44f-2e53162f4440; detaching it from the instance and deleting it from the info cache [ 915.594433] env[61911]: DEBUG nova.network.neutron [req-98e973dc-38a5-472f-8799-f799c9ef3eaa req-b31f463e-78e4-400c-acc5-b7a9b4eb67bf service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.828493] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.828493] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eaa35bb2-31c4-401d-9940-0bc43caf639e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.837036] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 915.837036] env[61911]: value = "task-1251308" [ 915.837036] env[61911]: _type = "Task" [ 915.837036] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.848199] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 915.848520] env[61911]: DEBUG nova.compute.manager [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 915.849612] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae21dc4-36b9-4816-a70b-14417e959d8d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.860051] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.860051] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.860245] env[61911]: DEBUG nova.network.neutron [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.905592] env[61911]: DEBUG nova.compute.manager [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 915.905922] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 915.907308] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad270ed-1d4e-4fa2-aca6-5ff80c5ce814 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.916145] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.916390] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7436bc5-bd3b-45cb-a9e2-f5f7befbb566 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.922137] env[61911]: DEBUG oslo_vmware.api [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 915.922137] env[61911]: value = "task-1251309" [ 915.922137] env[61911]: _type = "Task" [ 915.922137] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.929377] env[61911]: DEBUG oslo_vmware.api [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251309, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.938442] env[61911]: DEBUG nova.network.neutron [-] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.986379] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f6d5c5-00bf-a6cf-4db8-5a4fcde6dbc8, 'name': SearchDatastore_Task, 'duration_secs': 0.009533} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.986830] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.987133] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.987420] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.987619] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.987849] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.988156] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7469adf5-4508-4135-8013-ee8f2dadac31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.996460] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.996655] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.999610] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0939d30-a6e5-4a5e-8f41-fef1cd44c298 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.005362] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 916.005362] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f4492b-fa8b-c127-7dcd-8975abbb4276" [ 916.005362] env[61911]: _type = "Task" [ 916.005362] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.013282] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f4492b-fa8b-c127-7dcd-8975abbb4276, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.072993] env[61911]: DEBUG nova.network.neutron [-] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.101153] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36ba8c09-55df-46ee-b103-a6edc9e864c9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.110586] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c782bed6-26df-46a8-b399-7c388777c0cd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.125875] env[61911]: DEBUG nova.network.neutron [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Updated VIF entry in instance network info cache for port 4da57eb9-f431-463c-b1fc-292ec04a8cac. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.126264] env[61911]: DEBUG nova.network.neutron [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Updating instance_info_cache with network_info: [{"id": "4da57eb9-f431-463c-b1fc-292ec04a8cac", "address": "fa:16:3e:20:82:03", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4da57eb9-f4", "ovs_interfaceid": "4da57eb9-f431-463c-b1fc-292ec04a8cac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.140146] env[61911]: DEBUG nova.compute.manager [req-98e973dc-38a5-472f-8799-f799c9ef3eaa req-b31f463e-78e4-400c-acc5-b7a9b4eb67bf service nova] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Detach interface failed, port_id=3a287a5b-935a-49ef-b44f-2e53162f4440, reason: Instance b8cd7073-c712-4470-8ece-d144e8ce275f could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 916.432659] env[61911]: DEBUG oslo_vmware.api [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251309, 'name': PowerOffVM_Task, 'duration_secs': 0.21275} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.432659] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.432841] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.433080] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29c1a9aa-e5b4-49b7-a019-2ca64685c7bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.441031] env[61911]: INFO nova.compute.manager [-] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Took 1.32 seconds to deallocate network for instance. [ 916.498928] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.499135] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.499333] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleting the datastore file [datastore2] e9833b95-4162-42ba-87a4-d4cc790ac8e5 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.499608] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc27c8c0-ae84-49eb-bb84-7d5c0b8b8784 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.505817] env[61911]: DEBUG oslo_vmware.api [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for the task: (returnval){ [ 916.505817] env[61911]: value = "task-1251311" [ 916.505817] env[61911]: _type = "Task" [ 916.505817] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.520901] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f4492b-fa8b-c127-7dcd-8975abbb4276, 'name': SearchDatastore_Task, 'duration_secs': 0.008329} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.524271] env[61911]: DEBUG oslo_vmware.api [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251311, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.524686] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dbd7303-a0a6-429a-b6d7-4e0b56360603 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.529272] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 916.529272] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c45bfc-f743-d163-971b-0d0e3659b4c0" [ 916.529272] env[61911]: _type = "Task" [ 916.529272] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.539042] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c45bfc-f743-d163-971b-0d0e3659b4c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.580109] env[61911]: INFO nova.compute.manager [-] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Took 1.43 seconds to deallocate network for instance. [ 916.628962] env[61911]: DEBUG oslo_concurrency.lockutils [req-08b71364-d24e-4778-8d0e-921d8576ffad req-95fd2de3-896a-45fb-9fc0-40f6a0b2cb15 service nova] Releasing lock "refresh_cache-75577b68-b012-43d5-abdb-bd5f54c249d5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.660634] env[61911]: DEBUG nova.network.neutron [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.752196] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d37d4d-c947-4eac-a9e4-a6504fa82942 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.760031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8df825-b7c7-465c-bb83-046673645419 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.799622] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a78933-65f6-4326-84a9-01d1b55d59cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.807767] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034a4f76-ff9b-4799-be5d-2ccae9065bad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.821208] env[61911]: DEBUG nova.compute.provider_tree [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.947217] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.017603] env[61911]: DEBUG oslo_vmware.api [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251311, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.038675] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c45bfc-f743-d163-971b-0d0e3659b4c0, 'name': SearchDatastore_Task, 'duration_secs': 0.008766} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.038982] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.039277] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 75577b68-b012-43d5-abdb-bd5f54c249d5/75577b68-b012-43d5-abdb-bd5f54c249d5.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.039549] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49c51dc5-e7fe-43f5-9b5b-84f6122bc561 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.046241] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 917.046241] env[61911]: value = "task-1251312" [ 917.046241] env[61911]: _type = "Task" [ 917.046241] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.054052] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251312, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.089997] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.162954] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.324844] env[61911]: DEBUG nova.scheduler.client.report [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 917.335063] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.335357] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.335614] env[61911]: DEBUG nova.compute.manager [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Going to confirm migration 1 {{(pid=61911) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 917.391064] env[61911]: DEBUG nova.compute.manager [req-ccaf8d12-f60a-4a2b-a615-c81350f62d20 req-0818349a-0487-4662-9c5e-451ce0618b4a service nova] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Received event network-vif-deleted-2ce1966f-9d77-419b-8508-acc03ab5d271 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 917.445424] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.446387] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db353691-d6fa-4be4-bac0-23c7acd9e614 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.455762] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 917.456067] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d24b9c44-f4d9-48ad-b5e5-9994eed9c5b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.518667] env[61911]: DEBUG oslo_vmware.api [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Task: {'id': task-1251311, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.523168} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.519022] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.519276] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.519492] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.519672] env[61911]: INFO nova.compute.manager [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Took 1.61 seconds to destroy the instance on the hypervisor. [ 917.519971] env[61911]: DEBUG oslo.service.loopingcall [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.520502] env[61911]: DEBUG nova.compute.manager [-] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 917.520604] env[61911]: DEBUG nova.network.neutron [-] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.526398] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.526688] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.526787] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleting the datastore file [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.527049] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c3f4d5d-01fe-4ed2-ba70-985b7aff0c41 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.532786] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 917.532786] env[61911]: value = "task-1251314" [ 917.532786] env[61911]: _type = "Task" [ 917.532786] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.540941] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251314, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.554887] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251312, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436449} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.555149] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 75577b68-b012-43d5-abdb-bd5f54c249d5/75577b68-b012-43d5-abdb-bd5f54c249d5.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.555372] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.555630] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a69d5323-7d27-40d5-9ea8-08768d6f6b83 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.561888] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 917.561888] env[61911]: value = "task-1251315" [ 917.561888] env[61911]: _type = "Task" [ 917.561888] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.570063] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251315, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.686469] env[61911]: DEBUG nova.compute.manager [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received event network-vif-unplugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 917.686614] env[61911]: DEBUG oslo_concurrency.lockutils [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.686881] env[61911]: DEBUG oslo_concurrency.lockutils [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.687161] env[61911]: DEBUG oslo_concurrency.lockutils [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.687355] env[61911]: DEBUG nova.compute.manager [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] No waiting events found dispatching network-vif-unplugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 917.687531] env[61911]: WARNING nova.compute.manager [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received unexpected event network-vif-unplugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c for instance with vm_state shelved and task_state shelving_offloading. [ 917.687695] env[61911]: DEBUG nova.compute.manager [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received event network-changed-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 917.687855] env[61911]: DEBUG nova.compute.manager [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Refreshing instance network info cache due to event network-changed-f6d1cfde-6743-4e5f-ac24-8b71f63c293c. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 917.688074] env[61911]: DEBUG oslo_concurrency.lockutils [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.688266] env[61911]: DEBUG oslo_concurrency.lockutils [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.688434] env[61911]: DEBUG nova.network.neutron [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Refreshing network info cache for port f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.834251] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.834608] env[61911]: DEBUG nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 917.837120] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.890s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.837311] env[61911]: DEBUG nova.objects.instance [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lazy-loading 'resources' on Instance uuid 613544bc-92b9-4c8f-8a7b-f02205f347ff {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.894715] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.894921] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.895164] env[61911]: DEBUG nova.network.neutron [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.895368] env[61911]: DEBUG nova.objects.instance [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lazy-loading 'info_cache' on Instance uuid a35a11f1-3d4b-439e-9517-fa5a9f43af60 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.043193] env[61911]: DEBUG oslo_vmware.api [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251314, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146726} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.043487] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.043681] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.043865] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.067727] env[61911]: INFO nova.scheduler.client.report [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleted allocations for instance 8051e8ec-6d82-416c-858a-f0cdd00f869d [ 918.074581] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251315, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063485} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.074847] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.075669] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0ccc36-8607-4ebc-8ba5-cbab348aaebf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.097075] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 75577b68-b012-43d5-abdb-bd5f54c249d5/75577b68-b012-43d5-abdb-bd5f54c249d5.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.097392] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3082517-453f-4b2e-b893-14d953342e61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.117998] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 918.117998] env[61911]: value = "task-1251316" [ 918.117998] env[61911]: _type = "Task" [ 918.117998] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.126180] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251316, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.311969] env[61911]: DEBUG nova.network.neutron [-] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.340216] env[61911]: DEBUG nova.compute.utils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.347764] env[61911]: DEBUG nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 918.348105] env[61911]: DEBUG nova.network.neutron [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 918.395585] env[61911]: DEBUG nova.policy [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ac070f23f4044c19b5fbf5c70065298e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a74516eb1b73407ab23b68a0b56e6d7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.454427] env[61911]: DEBUG nova.network.neutron [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updated VIF entry in instance network info cache for port f6d1cfde-6743-4e5f-ac24-8b71f63c293c. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.454786] env[61911]: DEBUG nova.network.neutron [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": null, "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.551536] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d14d6fc-f03d-4b0c-97be-ef4941f2b057 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.559039] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f29f24-75d3-4c5a-8b14-4e2e20d0883f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.591084] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.592189] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82acbfb4-a1cb-4780-87ed-4a596b6c95f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.600244] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30f6e19-3fef-4d36-9d23-447b5a9784fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.613778] env[61911]: DEBUG nova.compute.provider_tree [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.626419] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251316, 'name': ReconfigVM_Task, 'duration_secs': 0.276492} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.626711] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 75577b68-b012-43d5-abdb-bd5f54c249d5/75577b68-b012-43d5-abdb-bd5f54c249d5.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.627395] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1cb9365-c00b-4bcc-a750-a05d4bda71a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.633979] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 918.633979] env[61911]: value = "task-1251317" [ 918.633979] env[61911]: _type = "Task" [ 918.633979] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.643227] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251317, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.644033] env[61911]: DEBUG nova.network.neutron [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Successfully created port: 7ba7fa67-0d47-400e-b536-623cccd0b119 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.816742] env[61911]: INFO nova.compute.manager [-] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Took 1.29 seconds to deallocate network for instance. [ 918.846290] env[61911]: DEBUG nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 918.957273] env[61911]: DEBUG oslo_concurrency.lockutils [req-5981cb2c-d519-42bd-9814-20448d9ec98e req-2c9a08b0-e56b-43c6-94e7-f7afffe61ce6 service nova] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.100508] env[61911]: DEBUG nova.network.neutron [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance_info_cache with network_info: [{"id": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "address": "fa:16:3e:c2:34:3f", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7ff300-bd", "ovs_interfaceid": "4e7ff300-bdd2-43d6-9fe5-646ffb508e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.116641] env[61911]: DEBUG nova.scheduler.client.report [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 919.143905] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251317, 'name': Rename_Task, 'duration_secs': 0.204936} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.144345] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.144454] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-841846de-f2b3-4159-8700-cfd24e5f7bcf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.151441] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 919.151441] env[61911]: value = "task-1251318" [ 919.151441] env[61911]: _type = "Task" [ 919.151441] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.160843] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251318, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.323780] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.603803] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-a35a11f1-3d4b-439e-9517-fa5a9f43af60" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.604126] env[61911]: DEBUG nova.objects.instance [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lazy-loading 'migration_context' on Instance uuid a35a11f1-3d4b-439e-9517-fa5a9f43af60 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.621187] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.623505] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.534s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.623505] env[61911]: DEBUG nova.objects.instance [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lazy-loading 'resources' on Instance uuid b8cd7073-c712-4470-8ece-d144e8ce275f {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.643106] env[61911]: INFO nova.scheduler.client.report [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted allocations for instance 613544bc-92b9-4c8f-8a7b-f02205f347ff [ 919.661730] env[61911]: DEBUG oslo_vmware.api [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251318, 'name': PowerOnVM_Task, 'duration_secs': 0.447646} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.662186] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.662273] env[61911]: INFO nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Took 7.14 seconds to spawn the instance on the hypervisor. [ 919.662471] env[61911]: DEBUG nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 919.663275] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f19b49-5f7b-46de-9e78-5ab4437cc216 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.711823] env[61911]: DEBUG nova.compute.manager [req-7e3edd5d-6590-4ced-bfd5-65bee1773bea req-5181e474-20f2-4194-b5bd-9cd79fc42f5c service nova] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Received event network-vif-deleted-c0501390-f2c7-4b6d-bd50-5d3a03c0e130 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 919.856461] env[61911]: DEBUG nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 919.877368] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 919.877622] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.877909] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 919.877984] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.878137] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 919.878294] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 919.878536] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 919.878830] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 919.878950] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 919.879356] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 919.879583] env[61911]: DEBUG nova.virt.hardware [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 919.880611] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09642da4-ec86-4bbf-adbb-098f8e5adc46 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.889554] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de1d7a4-a8a3-4fde-92af-7e8b3ba5898f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.037319] env[61911]: DEBUG nova.compute.manager [req-f7c9d3af-4a12-492c-aaaf-ae3b675942d5 req-d2955ce1-66e3-479e-9447-de723201f0c9 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Received event network-vif-plugged-7ba7fa67-0d47-400e-b536-623cccd0b119 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 920.037319] env[61911]: DEBUG oslo_concurrency.lockutils [req-f7c9d3af-4a12-492c-aaaf-ae3b675942d5 req-d2955ce1-66e3-479e-9447-de723201f0c9 service nova] Acquiring lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.037319] env[61911]: DEBUG oslo_concurrency.lockutils [req-f7c9d3af-4a12-492c-aaaf-ae3b675942d5 req-d2955ce1-66e3-479e-9447-de723201f0c9 service nova] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.037319] env[61911]: DEBUG oslo_concurrency.lockutils [req-f7c9d3af-4a12-492c-aaaf-ae3b675942d5 req-d2955ce1-66e3-479e-9447-de723201f0c9 service nova] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.037319] env[61911]: DEBUG nova.compute.manager [req-f7c9d3af-4a12-492c-aaaf-ae3b675942d5 req-d2955ce1-66e3-479e-9447-de723201f0c9 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] No waiting events found dispatching network-vif-plugged-7ba7fa67-0d47-400e-b536-623cccd0b119 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 920.037319] env[61911]: WARNING nova.compute.manager [req-f7c9d3af-4a12-492c-aaaf-ae3b675942d5 req-d2955ce1-66e3-479e-9447-de723201f0c9 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Received unexpected event network-vif-plugged-7ba7fa67-0d47-400e-b536-623cccd0b119 for instance with vm_state building and task_state spawning. [ 920.111062] env[61911]: DEBUG nova.objects.base [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 920.111062] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85e869a-61d6-4a42-acfc-e20bc0e37fdb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.134355] env[61911]: DEBUG nova.network.neutron [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Successfully updated port: 7ba7fa67-0d47-400e-b536-623cccd0b119 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.135925] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53a9896d-cf9d-4e7c-9458-cb82a84983dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.141814] env[61911]: DEBUG oslo_vmware.api [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 920.141814] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527ce6c1-233b-5ce4-87b8-7b2cbd25c7ac" [ 920.141814] env[61911]: _type = "Task" [ 920.141814] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.155216] env[61911]: DEBUG oslo_vmware.api [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527ce6c1-233b-5ce4-87b8-7b2cbd25c7ac, 'name': SearchDatastore_Task, 'duration_secs': 0.006912} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.155772] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1a5301b0-cab2-475f-825c-d05fc2b0f2b4 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "613544bc-92b9-4c8f-8a7b-f02205f347ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.709s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.156998] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.181188] env[61911]: INFO nova.compute.manager [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Took 11.75 seconds to build instance. [ 920.320816] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbbf274-f654-4341-9299-dd80e1a92707 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.328662] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "267075d7-c3d8-4520-b7d9-67504097f823" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.328935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "267075d7-c3d8-4520-b7d9-67504097f823" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.329224] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "267075d7-c3d8-4520-b7d9-67504097f823-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.329491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "267075d7-c3d8-4520-b7d9-67504097f823-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.329709] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "267075d7-c3d8-4520-b7d9-67504097f823-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.332121] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec380e5f-32de-416a-baf2-3363f658e7af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.335807] env[61911]: INFO nova.compute.manager [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Terminating instance [ 920.368580] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2255800-53f8-4081-b289-3023e1e5adf0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.376478] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d859e4df-5ad2-4847-8c64-15a8df0509fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.390222] env[61911]: DEBUG nova.compute.provider_tree [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.639472] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "refresh_cache-0ebf8d34-54c0-453e-b2e2-820feed6ee61" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.639730] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquired lock "refresh_cache-0ebf8d34-54c0-453e-b2e2-820feed6ee61" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.639901] env[61911]: DEBUG nova.network.neutron [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.682604] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e41fd4b1-974b-4b0c-a80c-8a06630326ee tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.265s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.748799] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.842028] env[61911]: DEBUG nova.compute.manager [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 920.842028] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.842028] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b11107c-9772-4dca-8b35-c6961ac7652a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.851020] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.851020] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6415578e-d700-4047-858c-0c109f5463e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.859413] env[61911]: DEBUG oslo_vmware.api [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 920.859413] env[61911]: value = "task-1251319" [ 920.859413] env[61911]: _type = "Task" [ 920.859413] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.868697] env[61911]: DEBUG oslo_vmware.api [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.892920] env[61911]: DEBUG nova.scheduler.client.report [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 921.169653] env[61911]: DEBUG nova.network.neutron [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.292370] env[61911]: DEBUG nova.network.neutron [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Updating instance_info_cache with network_info: [{"id": "7ba7fa67-0d47-400e-b536-623cccd0b119", "address": "fa:16:3e:65:79:c7", "network": {"id": "db43fd56-245d-4ff6-9001-726673f8b74c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1561147640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a74516eb1b73407ab23b68a0b56e6d7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9d50784-eb90-48ae-a4ea-2125c52a50d7", "external-id": "nsx-vlan-transportzone-657", "segmentation_id": 657, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ba7fa67-0d", "ovs_interfaceid": "7ba7fa67-0d47-400e-b536-623cccd0b119", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.371999] env[61911]: DEBUG oslo_vmware.api [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251319, 'name': PowerOffVM_Task, 'duration_secs': 0.245878} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.372363] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.372542] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.372798] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ff7a9d2-e0aa-43d5-90b6-c5e3623cf0cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.398083] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.400760] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.810s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.401013] env[61911]: DEBUG nova.objects.instance [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lazy-loading 'resources' on Instance uuid 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.415321] env[61911]: INFO nova.scheduler.client.report [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted allocations for instance b8cd7073-c712-4470-8ece-d144e8ce275f [ 921.456150] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.456591] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.456755] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleting the datastore file [datastore1] 267075d7-c3d8-4520-b7d9-67504097f823 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.457340] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9cc28b8-5476-4a1b-b081-2810761e9969 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.467568] env[61911]: DEBUG oslo_vmware.api [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 921.467568] env[61911]: value = "task-1251321" [ 921.467568] env[61911]: _type = "Task" [ 921.467568] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.480509] env[61911]: DEBUG oslo_vmware.api [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.619909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.619909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.795624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Releasing lock "refresh_cache-0ebf8d34-54c0-453e-b2e2-820feed6ee61" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.795960] env[61911]: DEBUG nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Instance network_info: |[{"id": "7ba7fa67-0d47-400e-b536-623cccd0b119", "address": "fa:16:3e:65:79:c7", "network": {"id": "db43fd56-245d-4ff6-9001-726673f8b74c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1561147640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a74516eb1b73407ab23b68a0b56e6d7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9d50784-eb90-48ae-a4ea-2125c52a50d7", "external-id": "nsx-vlan-transportzone-657", "segmentation_id": 657, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ba7fa67-0d", "ovs_interfaceid": "7ba7fa67-0d47-400e-b536-623cccd0b119", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 921.796575] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:79:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a9d50784-eb90-48ae-a4ea-2125c52a50d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ba7fa67-0d47-400e-b536-623cccd0b119', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.804266] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Creating folder: Project (a74516eb1b73407ab23b68a0b56e6d7e). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 921.804777] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aeb5c153-4819-432e-a5f9-a6007b4d7cc4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.815580] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Created folder: Project (a74516eb1b73407ab23b68a0b56e6d7e) in parent group-v269521. [ 921.815777] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Creating folder: Instances. Parent ref: group-v269625. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 921.816024] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f64f884-90b3-477c-b439-adac8f87d0a6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.823929] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Created folder: Instances in parent group-v269625. [ 921.824184] env[61911]: DEBUG oslo.service.loopingcall [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.824380] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.824581] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15c28321-abb6-4e14-94ee-4e3eba921600 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.842451] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.842451] env[61911]: value = "task-1251324" [ 921.842451] env[61911]: _type = "Task" [ 921.842451] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.849655] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251324, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.905245] env[61911]: DEBUG nova.objects.instance [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lazy-loading 'numa_topology' on Instance uuid 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.925140] env[61911]: DEBUG oslo_concurrency.lockutils [None req-02b24d0d-1be9-4ae3-a632-451fbea91394 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "b8cd7073-c712-4470-8ece-d144e8ce275f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.908s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.978544] env[61911]: DEBUG oslo_vmware.api [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251321, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134544} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.978828] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.979035] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.979258] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.979451] env[61911]: INFO nova.compute.manager [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Took 1.14 seconds to destroy the instance on the hypervisor. [ 921.979701] env[61911]: DEBUG oslo.service.loopingcall [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.979899] env[61911]: DEBUG nova.compute.manager [-] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 921.980091] env[61911]: DEBUG nova.network.neutron [-] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.062661] env[61911]: DEBUG nova.compute.manager [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Received event network-changed-7ba7fa67-0d47-400e-b536-623cccd0b119 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 922.062892] env[61911]: DEBUG nova.compute.manager [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Refreshing instance network info cache due to event network-changed-7ba7fa67-0d47-400e-b536-623cccd0b119. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 922.063137] env[61911]: DEBUG oslo_concurrency.lockutils [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] Acquiring lock "refresh_cache-0ebf8d34-54c0-453e-b2e2-820feed6ee61" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.063295] env[61911]: DEBUG oslo_concurrency.lockutils [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] Acquired lock "refresh_cache-0ebf8d34-54c0-453e-b2e2-820feed6ee61" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.063482] env[61911]: DEBUG nova.network.neutron [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Refreshing network info cache for port 7ba7fa67-0d47-400e-b536-623cccd0b119 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.122503] env[61911]: DEBUG nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 922.353471] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251324, 'name': CreateVM_Task, 'duration_secs': 0.386299} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.354128] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.354518] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.354748] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.355142] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.355443] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a048050-a87a-4f2e-a11b-2406eac06f32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.360679] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 922.360679] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5258615c-833e-9512-5809-d98bcf74002c" [ 922.360679] env[61911]: _type = "Task" [ 922.360679] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.364663] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.364754] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.364959] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.365190] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.365375] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.367728] env[61911]: INFO nova.compute.manager [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Terminating instance [ 922.374078] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5258615c-833e-9512-5809-d98bcf74002c, 'name': SearchDatastore_Task, 'duration_secs': 0.009508} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.374578] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.374834] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.375111] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.375282] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.375486] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.375777] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-baa13a66-2d1f-4054-813a-fd4db5d60932 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.385388] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.385388] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.386420] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43826659-350a-46b6-b53c-0a4bac2288d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.392814] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 922.392814] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e0969f-8d3c-b394-347b-d0b88b6a6053" [ 922.392814] env[61911]: _type = "Task" [ 922.392814] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.401129] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e0969f-8d3c-b394-347b-d0b88b6a6053, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.406937] env[61911]: DEBUG nova.objects.base [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Object Instance<8051e8ec-6d82-416c-858a-f0cdd00f869d> lazy-loaded attributes: resources,numa_topology {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 922.610419] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e015fa79-0b84-43e9-8dc2-2fbbc125c682 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.618499] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863a9b70-6daa-4334-84cf-5e8b61855108 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.653453] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0887fba8-d612-4cd7-8535-581e4e113fda {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.662849] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f60d5d5-f8de-4b0f-bfbb-4091f34fd06e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.668066] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.677830] env[61911]: DEBUG nova.compute.provider_tree [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.744018] env[61911]: DEBUG nova.network.neutron [-] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.800493] env[61911]: DEBUG nova.network.neutron [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Updated VIF entry in instance network info cache for port 7ba7fa67-0d47-400e-b536-623cccd0b119. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.800873] env[61911]: DEBUG nova.network.neutron [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Updating instance_info_cache with network_info: [{"id": "7ba7fa67-0d47-400e-b536-623cccd0b119", "address": "fa:16:3e:65:79:c7", "network": {"id": "db43fd56-245d-4ff6-9001-726673f8b74c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1561147640-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a74516eb1b73407ab23b68a0b56e6d7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a9d50784-eb90-48ae-a4ea-2125c52a50d7", "external-id": "nsx-vlan-transportzone-657", "segmentation_id": 657, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ba7fa67-0d", "ovs_interfaceid": "7ba7fa67-0d47-400e-b536-623cccd0b119", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.875499] env[61911]: DEBUG nova.compute.manager [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 922.875836] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.876859] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81be1667-3ded-46a0-9c0e-921d3b19225f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.884513] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.884756] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36df0b38-3d0b-466c-aa39-a812cc596d8d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.891592] env[61911]: DEBUG oslo_vmware.api [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 922.891592] env[61911]: value = "task-1251325" [ 922.891592] env[61911]: _type = "Task" [ 922.891592] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.902657] env[61911]: DEBUG oslo_vmware.api [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.906297] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e0969f-8d3c-b394-347b-d0b88b6a6053, 'name': SearchDatastore_Task, 'duration_secs': 0.010231} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.907168] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6ee68dc-0bfc-4123-8396-0edfb1271d83 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.912366] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 922.912366] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52be4e3b-76a9-4df5-2456-6755e129102b" [ 922.912366] env[61911]: _type = "Task" [ 922.912366] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.924144] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.924385] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.925451] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52be4e3b-76a9-4df5-2456-6755e129102b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.183325] env[61911]: DEBUG nova.scheduler.client.report [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 923.246376] env[61911]: INFO nova.compute.manager [-] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Took 1.27 seconds to deallocate network for instance. [ 923.304146] env[61911]: DEBUG oslo_concurrency.lockutils [req-85ce97ca-4c16-47b9-bd5d-b5542ca31815 req-5c24f25e-b32d-4aaf-afe5-f1e9d1a92707 service nova] Releasing lock "refresh_cache-0ebf8d34-54c0-453e-b2e2-820feed6ee61" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.401929] env[61911]: DEBUG oslo_vmware.api [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251325, 'name': PowerOffVM_Task, 'duration_secs': 0.172515} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.402224] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 923.402403] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 923.402659] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53bcfe4b-18e0-4558-844c-28f6b93d5f48 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.421646] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52be4e3b-76a9-4df5-2456-6755e129102b, 'name': SearchDatastore_Task, 'duration_secs': 0.009522} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.421935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.422236] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 0ebf8d34-54c0-453e-b2e2-820feed6ee61/0ebf8d34-54c0-453e-b2e2-820feed6ee61.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.422491] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-990c5709-6240-400d-b361-c4b79a96c7ad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.426779] env[61911]: DEBUG nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 923.430473] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 923.430473] env[61911]: value = "task-1251327" [ 923.430473] env[61911]: _type = "Task" [ 923.430473] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.438354] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251327, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.688189] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.287s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.691277] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.368s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.691531] env[61911]: DEBUG nova.objects.instance [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lazy-loading 'resources' on Instance uuid e9833b95-4162-42ba-87a4-d4cc790ac8e5 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.752747] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.831751] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "5b7d09e1-ee14-4001-8d73-14e763402670" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.832101] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.950723] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251327, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.957636] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.052838] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.053055] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.053203] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleting the datastore file [datastore1] f0dcfbbf-98ab-4e9f-838f-d9601d2bb045 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.053496] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c0435be0-4da6-4b64-945e-1ea7aafb9976 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.060711] env[61911]: DEBUG oslo_vmware.api [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 924.060711] env[61911]: value = "task-1251328" [ 924.060711] env[61911]: _type = "Task" [ 924.060711] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.069229] env[61911]: DEBUG oslo_vmware.api [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251328, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.155733] env[61911]: DEBUG nova.compute.manager [req-fce1934b-108b-456d-955b-ed253d43a5cb req-b55fcc40-201e-4915-974e-63c2b6cf7b60 service nova] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Received event network-vif-deleted-2e5e758d-6c28-4301-b6ec-99bd3a385c91 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 924.199886] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e1769fe-b0fe-417d-a111-3bbbb1d7c43e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.301s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.200855] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 3.452s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.201061] env[61911]: INFO nova.compute.manager [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Unshelving [ 924.336297] env[61911]: DEBUG nova.compute.utils [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.406091] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d62903-2fc7-449f-b735-05e9b8b4fe16 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.413618] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19224978-985d-4216-bd05-debe31ffbe04 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.446305] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e3c17e-61db-486b-94d0-fe16fc95f9ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.455449] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251327, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.788119} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.457510] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 0ebf8d34-54c0-453e-b2e2-820feed6ee61/0ebf8d34-54c0-453e-b2e2-820feed6ee61.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.457735] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.458034] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0086db2e-9620-4f14-890b-007db3daee17 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.460750] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20297330-0e1a-4806-bdc6-b59164441c4c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.476390] env[61911]: DEBUG nova.compute.provider_tree [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.479165] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 924.479165] env[61911]: value = "task-1251329" [ 924.479165] env[61911]: _type = "Task" [ 924.479165] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.487134] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251329, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.570202] env[61911]: DEBUG oslo_vmware.api [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251328, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.839669] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.980686] env[61911]: DEBUG nova.scheduler.client.report [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 924.993236] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251329, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075016} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.993533] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 924.994561] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651f27c5-b5b7-48a2-b060-15a61db579ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.019768] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 0ebf8d34-54c0-453e-b2e2-820feed6ee61/0ebf8d34-54c0-453e-b2e2-820feed6ee61.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.020345] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98e13654-4434-4ec0-b2bb-3ef37cc7aaf0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.039758] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 925.039758] env[61911]: value = "task-1251330" [ 925.039758] env[61911]: _type = "Task" [ 925.039758] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.047521] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251330, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.071083] env[61911]: DEBUG oslo_vmware.api [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251328, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.646085} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.071357] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.071614] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.071817] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.072020] env[61911]: INFO nova.compute.manager [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Took 2.20 seconds to destroy the instance on the hypervisor. [ 925.072266] env[61911]: DEBUG oslo.service.loopingcall [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.072498] env[61911]: DEBUG nova.compute.manager [-] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 925.072598] env[61911]: DEBUG nova.network.neutron [-] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.227185] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.489609] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.798s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.493074] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.335s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.514560] env[61911]: INFO nova.scheduler.client.report [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Deleted allocations for instance e9833b95-4162-42ba-87a4-d4cc790ac8e5 [ 925.550605] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251330, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.799218] env[61911]: DEBUG nova.network.neutron [-] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.896740] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "5b7d09e1-ee14-4001-8d73-14e763402670" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.896740] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.896740] env[61911]: INFO nova.compute.manager [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Attaching volume d4fa1f27-0c16-4b2d-9915-9fb3f46332b3 to /dev/sdb [ 925.926769] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9defa2e1-3c14-4f06-9bbe-aea6ec8dba20 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.934275] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c8a620-556a-4e5b-aa50-1760d5a18312 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.947557] env[61911]: DEBUG nova.virt.block_device [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updating existing volume attachment record: 042fcf2d-f484-43f4-b273-025704282710 {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 926.022414] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0f05cca6-011d-4bb3-ae2f-fc99501209d2 tempest-AttachInterfacesTestJSON-1828928692 tempest-AttachInterfacesTestJSON-1828928692-project-member] Lock "e9833b95-4162-42ba-87a4-d4cc790ac8e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.626s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.050049] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251330, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.184012] env[61911]: DEBUG nova.compute.manager [req-0a7f1806-cfed-4582-bb9d-2fdd9e1d25c0 req-187ac1ce-b833-4d35-a799-f4d2d90a9912 service nova] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Received event network-vif-deleted-8548c265-98c0-4ef3-80a4-3e6d5b65516b {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 926.193865] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3179ef-84fd-47b7-b2ec-158263ffc634 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.201572] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cd3b44-c520-4b9f-8f56-b35392d6ba1a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.233521] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13975965-32a1-4504-a93c-7c15b673f22b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.241319] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d9a38a-7235-4b98-b360-ec9588c21e62 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.260793] env[61911]: DEBUG nova.compute.provider_tree [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.301947] env[61911]: INFO nova.compute.manager [-] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Took 1.23 seconds to deallocate network for instance. [ 926.552159] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251330, 'name': ReconfigVM_Task, 'duration_secs': 1.106141} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.552373] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 0ebf8d34-54c0-453e-b2e2-820feed6ee61/0ebf8d34-54c0-453e-b2e2-820feed6ee61.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.553047] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf53f0b3-a970-41b6-a8f8-d8e182c3457e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.559848] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 926.559848] env[61911]: value = "task-1251334" [ 926.559848] env[61911]: _type = "Task" [ 926.559848] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.568636] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251334, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.761298] env[61911]: DEBUG nova.scheduler.client.report [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 926.809091] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.070034] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251334, 'name': Rename_Task, 'duration_secs': 0.133204} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.072778] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.072778] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14ad4ae2-5d7e-4fc0-b8d2-0dac1fa5a1c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.080169] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 927.080169] env[61911]: value = "task-1251335" [ 927.080169] env[61911]: _type = "Task" [ 927.080169] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.085375] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.589153] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251335, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.773163] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.280s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.775133] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.107s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.776640] env[61911]: INFO nova.compute.claims [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.091732] env[61911]: DEBUG oslo_vmware.api [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251335, 'name': PowerOnVM_Task, 'duration_secs': 0.633572} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.092068] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.092284] env[61911]: INFO nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Took 8.24 seconds to spawn the instance on the hypervisor. [ 928.092467] env[61911]: DEBUG nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 928.093280] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288c11f9-84a6-42b2-a208-dd7e0078deef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.363158] env[61911]: INFO nova.scheduler.client.report [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted allocation for migration 637723a6-e295-453e-8eae-95ce6f45696e [ 928.615687] env[61911]: INFO nova.compute.manager [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Took 13.12 seconds to build instance. [ 928.871290] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b3085bf1-bd92-40e6-8739-267c35a661e2 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.536s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.973915] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7290b385-98cf-433e-9133-5b63f6ff0e4e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.981906] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d24dcd-1754-4d5c-b3cf-30232b941567 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.013507] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31861b6-da0f-48dc-be9f-50b8baad85fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.021714] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62541722-602a-484a-bd41-1e4308d99938 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.035217] env[61911]: DEBUG nova.compute.provider_tree [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.046361] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.118269] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71373e0e-1c2f-46a2-b9ca-6cf9e9dd9ca6 tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.628s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.118596] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.072s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.118842] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.119052] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.119232] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.123081] env[61911]: INFO nova.compute.manager [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Terminating instance [ 929.478340] env[61911]: DEBUG oslo_concurrency.lockutils [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.478624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.538312] env[61911]: DEBUG nova.scheduler.client.report [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 929.627554] env[61911]: DEBUG nova.compute.manager [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 929.627820] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.628733] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beca085d-0405-49f2-a1d1-f8484191d77a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.636765] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.636997] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52687c52-0847-4e73-aefb-3c5d33fd6974 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.643021] env[61911]: DEBUG oslo_vmware.api [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 929.643021] env[61911]: value = "task-1251337" [ 929.643021] env[61911]: _type = "Task" [ 929.643021] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.650739] env[61911]: DEBUG oslo_vmware.api [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251337, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.753437] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.753804] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.754067] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.754284] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.754467] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.756673] env[61911]: INFO nova.compute.manager [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Terminating instance [ 929.981807] env[61911]: DEBUG nova.compute.utils [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.043508] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.044055] env[61911]: DEBUG nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 930.046718] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.294s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.046955] env[61911]: DEBUG nova.objects.instance [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lazy-loading 'resources' on Instance uuid 267075d7-c3d8-4520-b7d9-67504097f823 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.152382] env[61911]: DEBUG oslo_vmware.api [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251337, 'name': PowerOffVM_Task, 'duration_secs': 0.254919} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.152699] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.152852] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.153130] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7471ba9-cd3d-4e31-9ce6-eb206af1e578 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.240511] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.240798] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.241050] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Deleting the datastore file [datastore1] 0ebf8d34-54c0-453e-b2e2-820feed6ee61 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.241346] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b064705-4b10-4a5f-ac32-c687941007eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.248711] env[61911]: DEBUG oslo_vmware.api [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for the task: (returnval){ [ 930.248711] env[61911]: value = "task-1251339" [ 930.248711] env[61911]: _type = "Task" [ 930.248711] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.258288] env[61911]: DEBUG oslo_vmware.api [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.260806] env[61911]: DEBUG nova.compute.manager [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 930.261238] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.262102] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0849c0bf-4280-458c-bbf0-548079829fb0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.268951] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.269261] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef1c578c-7c75-4ff0-b53b-87736b8c8cab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.275161] env[61911]: DEBUG oslo_vmware.api [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 930.275161] env[61911]: value = "task-1251340" [ 930.275161] env[61911]: _type = "Task" [ 930.275161] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.286605] env[61911]: DEBUG oslo_vmware.api [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251340, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.440481] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f386424d-953d-431d-a56b-8542f2a3458c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.440978] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.484281] env[61911]: DEBUG oslo_concurrency.lockutils [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.493357] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 930.493591] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269629', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'name': 'volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b7d09e1-ee14-4001-8d73-14e763402670', 'attached_at': '', 'detached_at': '', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'serial': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 930.494491] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763cbb03-2fc3-43ec-9ac9-4d736d485ee7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.511927] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88f5025-0f51-4d22-b0b3-403ee81ef617 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.536328] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3/volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.536624] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d565b74-fe44-4a5d-9607-e177193ba57d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.550410] env[61911]: DEBUG nova.compute.utils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.551720] env[61911]: DEBUG nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 930.551965] env[61911]: DEBUG nova.network.neutron [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 930.561632] env[61911]: DEBUG oslo_vmware.api [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 930.561632] env[61911]: value = "task-1251341" [ 930.561632] env[61911]: _type = "Task" [ 930.561632] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.571535] env[61911]: DEBUG oslo_vmware.api [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251341, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.616246] env[61911]: DEBUG nova.policy [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe50c877fcef489cb798fafe73caaf8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813b3181120245e594809d096a621675', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.761070] env[61911]: DEBUG oslo_vmware.api [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Task: {'id': task-1251339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198983} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.764664] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.765217] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.765674] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.765909] env[61911]: INFO nova.compute.manager [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Took 1.14 seconds to destroy the instance on the hypervisor. [ 930.766240] env[61911]: DEBUG oslo.service.loopingcall [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.767167] env[61911]: DEBUG nova.compute.manager [-] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 930.767271] env[61911]: DEBUG nova.network.neutron [-] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 930.789252] env[61911]: DEBUG oslo_vmware.api [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251340, 'name': PowerOffVM_Task, 'duration_secs': 0.189098} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.789571] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.789760] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.790493] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b13ff672-0995-4bc5-af95-c8c8286d6c9d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.809200] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57e6df1-bd84-4c0b-9e26-27b21cbef36e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.815604] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42dc612-0f0a-4bc8-93b7-789832300330 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.852359] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939a4808-0f1a-46a0-9ae9-948cd47a8568 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.854885] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.855112] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.855304] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleting the datastore file [datastore1] a35a11f1-3d4b-439e-9517-fa5a9f43af60 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.855956] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1904be9-fea5-4896-9d7f-e9ca894f2db0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.860802] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2673c9e-2e1a-40da-8bce-d2ef99212c44 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.865795] env[61911]: DEBUG oslo_vmware.api [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 930.865795] env[61911]: value = "task-1251343" [ 930.865795] env[61911]: _type = "Task" [ 930.865795] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.877975] env[61911]: DEBUG nova.compute.provider_tree [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.889372] env[61911]: DEBUG oslo_vmware.api [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251343, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.931633] env[61911]: DEBUG nova.network.neutron [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Successfully created port: 9cdfed52-77e2-4eb9-9aac-f603cdea8d01 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 930.943373] env[61911]: DEBUG nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 931.056964] env[61911]: DEBUG nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 931.073056] env[61911]: DEBUG oslo_vmware.api [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251341, 'name': ReconfigVM_Task, 'duration_secs': 0.391623} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.073056] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Reconfigured VM instance instance-0000004b to attach disk [datastore1] volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3/volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.079332] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a338e47f-ed25-4488-846f-08d932736006 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.094808] env[61911]: DEBUG oslo_vmware.api [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 931.094808] env[61911]: value = "task-1251344" [ 931.094808] env[61911]: _type = "Task" [ 931.094808] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.104830] env[61911]: DEBUG oslo_vmware.api [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251344, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.380317] env[61911]: DEBUG oslo_vmware.api [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251343, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145097} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.381342] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.382016] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.382016] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.382320] env[61911]: INFO nova.compute.manager [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Took 1.12 seconds to destroy the instance on the hypervisor. [ 931.382779] env[61911]: DEBUG oslo.service.loopingcall [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.385170] env[61911]: DEBUG nova.compute.manager [-] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 931.385170] env[61911]: DEBUG nova.network.neutron [-] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.386378] env[61911]: DEBUG nova.scheduler.client.report [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 931.392478] env[61911]: DEBUG nova.compute.manager [req-58cc43bc-edac-428a-99dd-41dfa0a2fee0 req-22c0ed5f-4841-4688-8fab-d2fb663bf181 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Received event network-vif-deleted-7ba7fa67-0d47-400e-b536-623cccd0b119 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 931.392750] env[61911]: INFO nova.compute.manager [req-58cc43bc-edac-428a-99dd-41dfa0a2fee0 req-22c0ed5f-4841-4688-8fab-d2fb663bf181 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Neutron deleted interface 7ba7fa67-0d47-400e-b536-623cccd0b119; detaching it from the instance and deleting it from the info cache [ 931.392975] env[61911]: DEBUG nova.network.neutron [req-58cc43bc-edac-428a-99dd-41dfa0a2fee0 req-22c0ed5f-4841-4688-8fab-d2fb663bf181 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.468165] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.553186] env[61911]: DEBUG oslo_concurrency.lockutils [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.553477] env[61911]: DEBUG oslo_concurrency.lockutils [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.553732] env[61911]: INFO nova.compute.manager [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Attaching volume 8ade23e8-4f87-4c13-9e03-db70f4c4988d to /dev/sdb [ 931.613684] env[61911]: DEBUG oslo_vmware.api [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251344, 'name': ReconfigVM_Task, 'duration_secs': 0.140704} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.615119] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269629', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'name': 'volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b7d09e1-ee14-4001-8d73-14e763402670', 'attached_at': '', 'detached_at': '', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'serial': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 931.621584] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b106ec4f-09b6-4836-840d-cd1d6eca6c92 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.632867] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc4dc0e-ba02-4ddf-859e-0c465047e654 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.648612] env[61911]: DEBUG nova.virt.block_device [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating existing volume attachment record: 9269bb76-84bd-4473-bf38-4ffae2cb3cbc {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 931.862766] env[61911]: DEBUG nova.network.neutron [-] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.895472] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.898430] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.941s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.900452] env[61911]: INFO nova.compute.claims [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.903422] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80a535b3-b726-48f4-bdc8-5a95a46e04b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.913305] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f96865-f70d-4702-ae95-87e290e7aed1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.938762] env[61911]: INFO nova.scheduler.client.report [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted allocations for instance 267075d7-c3d8-4520-b7d9-67504097f823 [ 931.955674] env[61911]: DEBUG nova.compute.manager [req-58cc43bc-edac-428a-99dd-41dfa0a2fee0 req-22c0ed5f-4841-4688-8fab-d2fb663bf181 service nova] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Detach interface failed, port_id=7ba7fa67-0d47-400e-b536-623cccd0b119, reason: Instance 0ebf8d34-54c0-453e-b2e2-820feed6ee61 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 932.070429] env[61911]: DEBUG nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 932.092520] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 932.092652] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.092810] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 932.093403] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.093403] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 932.093512] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 932.093899] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 932.094148] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 932.094371] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 932.094649] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 932.095616] env[61911]: DEBUG nova.virt.hardware [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 932.095741] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4028a4f9-42ae-4417-9adb-9efaae928399 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.104948] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e15b00-3b3a-4853-9d64-6c5dc5cba0ec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.253043] env[61911]: DEBUG nova.network.neutron [-] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.365470] env[61911]: INFO nova.compute.manager [-] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Took 1.60 seconds to deallocate network for instance. [ 932.381364] env[61911]: DEBUG nova.compute.manager [req-d24a012d-6a8c-42b7-89f1-aaba295b93b7 req-3cffb857-599c-49d3-97c2-78c91ba0d64a service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Received event network-vif-plugged-9cdfed52-77e2-4eb9-9aac-f603cdea8d01 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 932.381619] env[61911]: DEBUG oslo_concurrency.lockutils [req-d24a012d-6a8c-42b7-89f1-aaba295b93b7 req-3cffb857-599c-49d3-97c2-78c91ba0d64a service nova] Acquiring lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.382718] env[61911]: DEBUG oslo_concurrency.lockutils [req-d24a012d-6a8c-42b7-89f1-aaba295b93b7 req-3cffb857-599c-49d3-97c2-78c91ba0d64a service nova] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.382784] env[61911]: DEBUG oslo_concurrency.lockutils [req-d24a012d-6a8c-42b7-89f1-aaba295b93b7 req-3cffb857-599c-49d3-97c2-78c91ba0d64a service nova] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.383117] env[61911]: DEBUG nova.compute.manager [req-d24a012d-6a8c-42b7-89f1-aaba295b93b7 req-3cffb857-599c-49d3-97c2-78c91ba0d64a service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] No waiting events found dispatching network-vif-plugged-9cdfed52-77e2-4eb9-9aac-f603cdea8d01 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 932.383241] env[61911]: WARNING nova.compute.manager [req-d24a012d-6a8c-42b7-89f1-aaba295b93b7 req-3cffb857-599c-49d3-97c2-78c91ba0d64a service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Received unexpected event network-vif-plugged-9cdfed52-77e2-4eb9-9aac-f603cdea8d01 for instance with vm_state building and task_state spawning. [ 932.449054] env[61911]: DEBUG oslo_concurrency.lockutils [None req-be627abd-a565-48b5-8527-026c18930fa3 tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "267075d7-c3d8-4520-b7d9-67504097f823" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.120s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.477045] env[61911]: DEBUG nova.network.neutron [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Successfully updated port: 9cdfed52-77e2-4eb9-9aac-f603cdea8d01 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.617026] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.618071] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.663444] env[61911]: DEBUG nova.objects.instance [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'flavor' on Instance uuid 5b7d09e1-ee14-4001-8d73-14e763402670 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.755467] env[61911]: INFO nova.compute.manager [-] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Took 1.37 seconds to deallocate network for instance. [ 932.850332] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "66472f43-537d-4eb3-8d49-d40627a8809d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.850646] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "66472f43-537d-4eb3-8d49-d40627a8809d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.850938] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "66472f43-537d-4eb3-8d49-d40627a8809d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.851090] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "66472f43-537d-4eb3-8d49-d40627a8809d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.851273] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "66472f43-537d-4eb3-8d49-d40627a8809d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.853768] env[61911]: INFO nova.compute.manager [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Terminating instance [ 932.872121] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.980208] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "refresh_cache-8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.980353] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "refresh_cache-8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.980504] env[61911]: DEBUG nova.network.neutron [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.120898] env[61911]: DEBUG nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 933.157201] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653b31d0-26bb-4a5b-9e98-a287538b5d53 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.168712] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5d4f35c3-dc2f-46d6-b86e-4e256cc30cdc tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.273s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.170346] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db426cc3-d92f-444f-8f59-a765b99fb3cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.204592] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843c8952-474a-4551-b6bc-10f8a438b152 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.213345] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e8bab4-a57f-4eeb-95cb-3b6b598f4640 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.227281] env[61911]: DEBUG nova.compute.provider_tree [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.262228] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.359068] env[61911]: DEBUG nova.compute.manager [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 933.359068] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.359357] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8973c77b-290e-41d7-a8cc-cf7f55d80a5e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.368312] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.368608] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6b744576-ebbc-466f-bb71-a6a399cd3c8e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.375043] env[61911]: DEBUG oslo_vmware.api [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 933.375043] env[61911]: value = "task-1251348" [ 933.375043] env[61911]: _type = "Task" [ 933.375043] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.383389] env[61911]: DEBUG oslo_vmware.api [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251348, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.460231] env[61911]: DEBUG nova.compute.manager [req-abba423f-2834-4805-9808-b22de95f7b8e req-7c6e50d0-e787-4654-ad34-f0582846ae24 service nova] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Received event network-vif-deleted-4e7ff300-bdd2-43d6-9fe5-646ffb508e5f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 933.529206] env[61911]: DEBUG nova.network.neutron [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.649564] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.730984] env[61911]: DEBUG nova.scheduler.client.report [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 933.834849] env[61911]: DEBUG nova.network.neutron [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Updating instance_info_cache with network_info: [{"id": "9cdfed52-77e2-4eb9-9aac-f603cdea8d01", "address": "fa:16:3e:7a:b8:86", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cdfed52-77", "ovs_interfaceid": "9cdfed52-77e2-4eb9-9aac-f603cdea8d01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.884610] env[61911]: DEBUG oslo_vmware.api [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251348, 'name': PowerOffVM_Task, 'duration_secs': 0.192474} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.884886] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.885074] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.885335] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d15272a0-0d07-4d61-a95d-d8fdef6801a6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.958479] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.958795] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.958896] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleting the datastore file [datastore2] 66472f43-537d-4eb3-8d49-d40627a8809d {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.959184] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-942c2ca9-09ff-43fe-8dff-c83fc688ddb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.966959] env[61911]: DEBUG oslo_vmware.api [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for the task: (returnval){ [ 933.966959] env[61911]: value = "task-1251351" [ 933.966959] env[61911]: _type = "Task" [ 933.966959] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.975323] env[61911]: DEBUG oslo_vmware.api [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251351, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.236033] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.236488] env[61911]: DEBUG nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 934.239321] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.012s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.239479] env[61911]: DEBUG nova.objects.instance [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lazy-loading 'pci_requests' on Instance uuid 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.337176] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "refresh_cache-8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.337530] env[61911]: DEBUG nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Instance network_info: |[{"id": "9cdfed52-77e2-4eb9-9aac-f603cdea8d01", "address": "fa:16:3e:7a:b8:86", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cdfed52-77", "ovs_interfaceid": "9cdfed52-77e2-4eb9-9aac-f603cdea8d01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 934.337996] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:b8:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9cdfed52-77e2-4eb9-9aac-f603cdea8d01', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.345525] env[61911]: DEBUG oslo.service.loopingcall [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.345748] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 934.345982] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6257fa26-9440-44d2-8207-c860893922a0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.365413] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.365413] env[61911]: value = "task-1251352" [ 934.365413] env[61911]: _type = "Task" [ 934.365413] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.372682] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251352, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.409861] env[61911]: DEBUG nova.compute.manager [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Received event network-changed-9cdfed52-77e2-4eb9-9aac-f603cdea8d01 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 934.410204] env[61911]: DEBUG nova.compute.manager [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Refreshing instance network info cache due to event network-changed-9cdfed52-77e2-4eb9-9aac-f603cdea8d01. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 934.410443] env[61911]: DEBUG oslo_concurrency.lockutils [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] Acquiring lock "refresh_cache-8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.410443] env[61911]: DEBUG oslo_concurrency.lockutils [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] Acquired lock "refresh_cache-8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.410604] env[61911]: DEBUG nova.network.neutron [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Refreshing network info cache for port 9cdfed52-77e2-4eb9-9aac-f603cdea8d01 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.477061] env[61911]: DEBUG oslo_vmware.api [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Task: {'id': task-1251351, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144973} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.478103] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.478103] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.478240] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.478343] env[61911]: INFO nova.compute.manager [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 934.478603] env[61911]: DEBUG oslo.service.loopingcall [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.478830] env[61911]: DEBUG nova.compute.manager [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 934.478921] env[61911]: DEBUG nova.network.neutron [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 934.645977] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "fddba36d-1b15-43fb-9e99-68880e8a235c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.646375] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.743348] env[61911]: DEBUG nova.compute.utils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.745919] env[61911]: DEBUG nova.objects.instance [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lazy-loading 'numa_topology' on Instance uuid 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.746818] env[61911]: DEBUG nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 934.746985] env[61911]: DEBUG nova.network.neutron [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.805006] env[61911]: DEBUG nova.policy [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d0056b50ef7429ea460ff4c1156d132', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf7a7083777a43a5bb9fc04584dcdd06', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.876103] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251352, 'name': CreateVM_Task, 'duration_secs': 0.367925} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.876292] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.877009] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.877195] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.881083] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.881370] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d01b093-d114-4721-b6c2-b3da953b05b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.886463] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 934.886463] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d20724-1e76-e24f-fbbd-effd97a9f27b" [ 934.886463] env[61911]: _type = "Task" [ 934.886463] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.894685] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d20724-1e76-e24f-fbbd-effd97a9f27b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.109662] env[61911]: DEBUG nova.network.neutron [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Updated VIF entry in instance network info cache for port 9cdfed52-77e2-4eb9-9aac-f603cdea8d01. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.110060] env[61911]: DEBUG nova.network.neutron [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Updating instance_info_cache with network_info: [{"id": "9cdfed52-77e2-4eb9-9aac-f603cdea8d01", "address": "fa:16:3e:7a:b8:86", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cdfed52-77", "ovs_interfaceid": "9cdfed52-77e2-4eb9-9aac-f603cdea8d01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.146036] env[61911]: DEBUG nova.network.neutron [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Successfully created port: 38e33f6e-614a-4041-a8d0-136e96470621 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.148474] env[61911]: DEBUG nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 935.253755] env[61911]: DEBUG nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 935.257158] env[61911]: INFO nova.compute.claims [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.287979] env[61911]: DEBUG nova.network.neutron [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.397045] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d20724-1e76-e24f-fbbd-effd97a9f27b, 'name': SearchDatastore_Task, 'duration_secs': 0.011734} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.397509] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.397630] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.397900] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.398067] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.398259] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.398509] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3fd6278c-e325-4793-b607-846d4c6cbbad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.406403] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.406580] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.407259] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d9c16b5-5456-40f4-880e-0ae672571749 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.412604] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 935.412604] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527d42d6-ca38-54f9-b439-fce3a8ad8a92" [ 935.412604] env[61911]: _type = "Task" [ 935.412604] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.419985] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527d42d6-ca38-54f9-b439-fce3a8ad8a92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.613946] env[61911]: DEBUG oslo_concurrency.lockutils [req-eab63e8d-db52-42ba-8dc3-26c7e9297cdb req-7ed29908-9419-4033-9b81-d35777873565 service nova] Releasing lock "refresh_cache-8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.671402] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.790921] env[61911]: INFO nova.compute.manager [-] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Took 1.31 seconds to deallocate network for instance. [ 935.922974] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527d42d6-ca38-54f9-b439-fce3a8ad8a92, 'name': SearchDatastore_Task, 'duration_secs': 0.010158} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.923716] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaedc60e-70e8-483c-8acb-b9919c29b639 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.928392] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 935.928392] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52cd8b88-cbd8-462d-49e5-65be2759f925" [ 935.928392] env[61911]: _type = "Task" [ 935.928392] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.935490] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52cd8b88-cbd8-462d-49e5-65be2759f925, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.205710] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 936.206112] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269631', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'name': 'volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c', 'attached_at': '', 'detached_at': '', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'serial': '8ade23e8-4f87-4c13-9e03-db70f4c4988d'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 936.206979] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa80bd7-a596-40e9-a5c4-25d65f019170 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.223744] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2242fd8-0fc2-4af8-9096-06570f9a3095 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.247955] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d/volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.248248] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cf981fa-9715-423e-867e-c01260fa0bdb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.266388] env[61911]: DEBUG nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 936.271982] env[61911]: DEBUG oslo_vmware.api [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 936.271982] env[61911]: value = "task-1251353" [ 936.271982] env[61911]: _type = "Task" [ 936.271982] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.282788] env[61911]: DEBUG oslo_vmware.api [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251353, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.292505] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 936.292755] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.292910] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 936.293120] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.293286] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 936.293440] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 936.293644] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 936.293806] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 936.293977] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 936.294158] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 936.294332] env[61911]: DEBUG nova.virt.hardware [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 936.295149] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c574a6a6-25ae-4e7d-8f55-a93ad2132e63 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.298164] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.304391] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9015c8d-2f13-48db-846b-5804c3eb1c09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.437499] env[61911]: DEBUG nova.compute.manager [req-108fe754-3cb9-423a-9a25-8ee02b1628d2 req-468544b3-4fc7-4edb-b203-f6798d130598 service nova] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Received event network-vif-deleted-90825141-c63e-4853-8f27-6553cc0f3ecc {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 936.442226] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52cd8b88-cbd8-462d-49e5-65be2759f925, 'name': SearchDatastore_Task, 'duration_secs': 0.008779} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.442391] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.442704] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a/8d9881fe-4bb5-4d85-9be9-32c2eca32a9a.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.442884] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5290ecb6-ba44-4484-840d-7418472c4cdb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.452097] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 936.452097] env[61911]: value = "task-1251354" [ 936.452097] env[61911]: _type = "Task" [ 936.452097] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.461196] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.479852] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aee8e4a-a8fa-46ef-89aa-479fa165c11c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.486871] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2db9fde-6019-4106-8872-876d1948fc42 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.516814] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4a3998-e095-4668-87c8-be91f9ae8318 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.524138] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f79c21-d824-4dff-9f29-7e0f3bef314d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.537845] env[61911]: DEBUG nova.compute.provider_tree [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.557778] env[61911]: DEBUG nova.compute.manager [req-5f09fa06-719b-4f23-8e91-e57d24047719 req-a68a6fab-e338-49cf-a6c6-9dbcf3bb6224 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Received event network-vif-plugged-38e33f6e-614a-4041-a8d0-136e96470621 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 936.558024] env[61911]: DEBUG oslo_concurrency.lockutils [req-5f09fa06-719b-4f23-8e91-e57d24047719 req-a68a6fab-e338-49cf-a6c6-9dbcf3bb6224 service nova] Acquiring lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.558277] env[61911]: DEBUG oslo_concurrency.lockutils [req-5f09fa06-719b-4f23-8e91-e57d24047719 req-a68a6fab-e338-49cf-a6c6-9dbcf3bb6224 service nova] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.558456] env[61911]: DEBUG oslo_concurrency.lockutils [req-5f09fa06-719b-4f23-8e91-e57d24047719 req-a68a6fab-e338-49cf-a6c6-9dbcf3bb6224 service nova] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.558634] env[61911]: DEBUG nova.compute.manager [req-5f09fa06-719b-4f23-8e91-e57d24047719 req-a68a6fab-e338-49cf-a6c6-9dbcf3bb6224 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] No waiting events found dispatching network-vif-plugged-38e33f6e-614a-4041-a8d0-136e96470621 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 936.558807] env[61911]: WARNING nova.compute.manager [req-5f09fa06-719b-4f23-8e91-e57d24047719 req-a68a6fab-e338-49cf-a6c6-9dbcf3bb6224 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Received unexpected event network-vif-plugged-38e33f6e-614a-4041-a8d0-136e96470621 for instance with vm_state building and task_state spawning. [ 936.670904] env[61911]: DEBUG nova.network.neutron [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Successfully updated port: 38e33f6e-614a-4041-a8d0-136e96470621 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.781983] env[61911]: DEBUG oslo_vmware.api [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251353, 'name': ReconfigVM_Task, 'duration_secs': 0.340674} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.782702] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfigured VM instance instance-0000004d to attach disk [datastore1] volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d/volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.787982] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9dfcaa8-d3f7-441f-a655-479e627e6740 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.805915] env[61911]: DEBUG oslo_vmware.api [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 936.805915] env[61911]: value = "task-1251355" [ 936.805915] env[61911]: _type = "Task" [ 936.805915] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.817199] env[61911]: DEBUG oslo_vmware.api [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251355, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.961439] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48951} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.961714] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a/8d9881fe-4bb5-4d85-9be9-32c2eca32a9a.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 936.961944] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.962253] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d136ea41-a4b5-40f4-9893-1f87984dd9c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.968651] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 936.968651] env[61911]: value = "task-1251356" [ 936.968651] env[61911]: _type = "Task" [ 936.968651] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.976653] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.042936] env[61911]: DEBUG nova.scheduler.client.report [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 937.173664] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "refresh_cache-3c4c8943-5324-4c3e-b7e0-7ea5972bc026" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.174491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "refresh_cache-3c4c8943-5324-4c3e-b7e0-7ea5972bc026" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.174491] env[61911]: DEBUG nova.network.neutron [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.317224] env[61911]: DEBUG oslo_vmware.api [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251355, 'name': ReconfigVM_Task, 'duration_secs': 0.22597} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.317567] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269631', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'name': 'volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c', 'attached_at': '', 'detached_at': '', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'serial': '8ade23e8-4f87-4c13-9e03-db70f4c4988d'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 937.477820] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06195} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.478160] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.478830] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4160a82f-a2b9-4657-8c23-da29b246715d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.500533] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a/8d9881fe-4bb5-4d85-9be9-32c2eca32a9a.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.500780] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23b7dccc-d9df-488b-b8fb-09124bf89533 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.519813] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 937.519813] env[61911]: value = "task-1251357" [ 937.519813] env[61911]: _type = "Task" [ 937.519813] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.529257] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251357, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.547477] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.308s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.549862] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.741s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.550194] env[61911]: DEBUG nova.objects.instance [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lazy-loading 'resources' on Instance uuid f0dcfbbf-98ab-4e9f-838f-d9601d2bb045 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.578380] env[61911]: INFO nova.network.neutron [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating port f6d1cfde-6743-4e5f-ac24-8b71f63c293c with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 937.718977] env[61911]: DEBUG nova.network.neutron [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.858602] env[61911]: DEBUG nova.network.neutron [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Updating instance_info_cache with network_info: [{"id": "38e33f6e-614a-4041-a8d0-136e96470621", "address": "fa:16:3e:47:e7:d5", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e33f6e-61", "ovs_interfaceid": "38e33f6e-614a-4041-a8d0-136e96470621", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.030492] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251357, 'name': ReconfigVM_Task, 'duration_secs': 0.267129} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.031132] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a/8d9881fe-4bb5-4d85-9be9-32c2eca32a9a.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.031918] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9b3bf80-e460-4144-a52b-f421c224a226 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.040017] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 938.040017] env[61911]: value = "task-1251358" [ 938.040017] env[61911]: _type = "Task" [ 938.040017] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.048755] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251358, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.258658] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6f0642-ce9e-487c-99c9-d035ea82874d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.266047] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7da48e3-ad1c-4002-8078-b87eafca95f4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.294840] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df8d81a-d548-4947-9b9a-3daaadb7c797 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.302094] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b631c8c-7808-416a-a543-00e88ff76927 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.314562] env[61911]: DEBUG nova.compute.provider_tree [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.355086] env[61911]: DEBUG nova.objects.instance [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lazy-loading 'flavor' on Instance uuid 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.361901] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "refresh_cache-3c4c8943-5324-4c3e-b7e0-7ea5972bc026" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.361901] env[61911]: DEBUG nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Instance network_info: |[{"id": "38e33f6e-614a-4041-a8d0-136e96470621", "address": "fa:16:3e:47:e7:d5", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e33f6e-61", "ovs_interfaceid": "38e33f6e-614a-4041-a8d0-136e96470621", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 938.362126] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:e7:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '38e33f6e-614a-4041-a8d0-136e96470621', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.369988] env[61911]: DEBUG oslo.service.loopingcall [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.370808] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.371080] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af4d3200-d8aa-486b-9619-1db23e6b30d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.390675] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.390675] env[61911]: value = "task-1251359" [ 938.390675] env[61911]: _type = "Task" [ 938.390675] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.399921] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251359, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.548094] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251358, 'name': Rename_Task, 'duration_secs': 0.136041} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.548572] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.548572] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c10877e-db62-4135-ad1d-42e3db32a05c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.554774] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 938.554774] env[61911]: value = "task-1251360" [ 938.554774] env[61911]: _type = "Task" [ 938.554774] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.561900] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.587261] env[61911]: DEBUG nova.compute.manager [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Received event network-changed-38e33f6e-614a-4041-a8d0-136e96470621 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 938.587455] env[61911]: DEBUG nova.compute.manager [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Refreshing instance network info cache due to event network-changed-38e33f6e-614a-4041-a8d0-136e96470621. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 938.587680] env[61911]: DEBUG oslo_concurrency.lockutils [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] Acquiring lock "refresh_cache-3c4c8943-5324-4c3e-b7e0-7ea5972bc026" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.588136] env[61911]: DEBUG oslo_concurrency.lockutils [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] Acquired lock "refresh_cache-3c4c8943-5324-4c3e-b7e0-7ea5972bc026" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.588136] env[61911]: DEBUG nova.network.neutron [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Refreshing network info cache for port 38e33f6e-614a-4041-a8d0-136e96470621 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.818649] env[61911]: DEBUG nova.scheduler.client.report [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 938.859902] env[61911]: DEBUG oslo_concurrency.lockutils [None req-21c6747e-3fab-494c-9351-4922d3a36dec tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.306s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.900742] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251359, 'name': CreateVM_Task, 'duration_secs': 0.423915} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.900915] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.901630] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.901809] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.902149] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.902394] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9727f96a-6692-4796-99cd-b39cc3b76ddc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.907008] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 938.907008] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5238660e-c4a9-f41e-f3ba-160c00247afe" [ 938.907008] env[61911]: _type = "Task" [ 938.907008] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.914660] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5238660e-c4a9-f41e-f3ba-160c00247afe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.973452] env[61911]: INFO nova.compute.manager [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Rescuing [ 938.973739] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.973896] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.974115] env[61911]: DEBUG nova.network.neutron [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.065131] env[61911]: DEBUG oslo_vmware.api [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251360, 'name': PowerOnVM_Task, 'duration_secs': 0.504412} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.065410] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.065621] env[61911]: INFO nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Took 6.99 seconds to spawn the instance on the hypervisor. [ 939.065810] env[61911]: DEBUG nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 939.066547] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771fb86e-91c8-447c-8561-e0e3908556f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.288206] env[61911]: DEBUG nova.network.neutron [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Updated VIF entry in instance network info cache for port 38e33f6e-614a-4041-a8d0-136e96470621. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.288579] env[61911]: DEBUG nova.network.neutron [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Updating instance_info_cache with network_info: [{"id": "38e33f6e-614a-4041-a8d0-136e96470621", "address": "fa:16:3e:47:e7:d5", "network": {"id": "e8379e5f-f220-4605-b1a9-c0b2ad7865d3", "bridge": "br-int", "label": "tempest-ImagesTestJSON-355809618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cf7a7083777a43a5bb9fc04584dcdd06", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38e33f6e-61", "ovs_interfaceid": "38e33f6e-614a-4041-a8d0-136e96470621", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.324652] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.326077] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.858s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.327576] env[61911]: INFO nova.compute.claims [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.345242] env[61911]: INFO nova.scheduler.client.report [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted allocations for instance f0dcfbbf-98ab-4e9f-838f-d9601d2bb045 [ 939.419166] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5238660e-c4a9-f41e-f3ba-160c00247afe, 'name': SearchDatastore_Task, 'duration_secs': 0.009588} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.419676] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.419942] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.420225] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.420384] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.420571] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.420842] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c2a2190-a022-4e74-b68e-cc0edd4beb6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.432533] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.432716] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.433462] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09ebca77-4303-498b-8175-c639dabfcf74 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.438722] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 939.438722] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5223fe9f-31c9-61ca-9237-c92a7ed2ce51" [ 939.438722] env[61911]: _type = "Task" [ 939.438722] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.446704] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5223fe9f-31c9-61ca-9237-c92a7ed2ce51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.479079] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.479211] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.479403] env[61911]: DEBUG nova.network.neutron [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.589477] env[61911]: INFO nova.compute.manager [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Took 16.94 seconds to build instance. [ 939.792023] env[61911]: DEBUG oslo_concurrency.lockutils [req-916af112-7805-46db-b361-a13779982910 req-a48268e7-14f8-4663-91c8-ac3050599927 service nova] Releasing lock "refresh_cache-3c4c8943-5324-4c3e-b7e0-7ea5972bc026" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.840696] env[61911]: DEBUG nova.network.neutron [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [{"id": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "address": "fa:16:3e:c5:e0:b6", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce08c90-28", "ovs_interfaceid": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.851886] env[61911]: DEBUG oslo_concurrency.lockutils [None req-5c66c2ef-8377-4f4c-805d-a7bb36c3b9b9 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "f0dcfbbf-98ab-4e9f-838f-d9601d2bb045" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.487s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.949176] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5223fe9f-31c9-61ca-9237-c92a7ed2ce51, 'name': SearchDatastore_Task, 'duration_secs': 0.012287} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.949952] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22ce1972-963d-4a3c-af7f-19200f84f366 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.954860] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 939.954860] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528f41c0-08b3-c81d-be63-3e9bd2ab2619" [ 939.954860] env[61911]: _type = "Task" [ 939.954860] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.961810] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528f41c0-08b3-c81d-be63-3e9bd2ab2619, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.093968] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d093e919-80bd-4dfd-b725-5072ec5f1d56 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.474s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.174347] env[61911]: DEBUG nova.network.neutron [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.312534] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.312823] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.313077] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.313288] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.313469] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.315970] env[61911]: INFO nova.compute.manager [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Terminating instance [ 940.342870] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.465287] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528f41c0-08b3-c81d-be63-3e9bd2ab2619, 'name': SearchDatastore_Task, 'duration_secs': 0.012657} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.467629] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.467905] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 3c4c8943-5324-4c3e-b7e0-7ea5972bc026/3c4c8943-5324-4c3e-b7e0-7ea5972bc026.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.468351] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ffe8135-aecc-480f-9f91-4ade52f7dbc7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.474226] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 940.474226] env[61911]: value = "task-1251361" [ 940.474226] env[61911]: _type = "Task" [ 940.474226] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.483951] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251361, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.529931] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7c7a50-a755-4882-8c1f-045f4a9c4c2d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.537220] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c6cbc5-e69c-4d29-92d7-9dfe75302b1d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.569557] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e95b7b-8668-46d9-b57b-877c6ac19071 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.576466] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845b423e-dd24-4ed6-b2ce-7b00e76ef117 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.589296] env[61911]: DEBUG nova.compute.provider_tree [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.612776] env[61911]: DEBUG nova.compute.manager [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received event network-vif-plugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 940.613076] env[61911]: DEBUG oslo_concurrency.lockutils [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.613239] env[61911]: DEBUG oslo_concurrency.lockutils [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.613414] env[61911]: DEBUG oslo_concurrency.lockutils [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.613684] env[61911]: DEBUG nova.compute.manager [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] No waiting events found dispatching network-vif-plugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 940.613778] env[61911]: WARNING nova.compute.manager [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received unexpected event network-vif-plugged-f6d1cfde-6743-4e5f-ac24-8b71f63c293c for instance with vm_state shelved_offloaded and task_state spawning. [ 940.613966] env[61911]: DEBUG nova.compute.manager [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received event network-changed-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 940.614346] env[61911]: DEBUG nova.compute.manager [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Refreshing instance network info cache due to event network-changed-f6d1cfde-6743-4e5f-ac24-8b71f63c293c. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 940.615129] env[61911]: DEBUG oslo_concurrency.lockutils [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.676700] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.679371] env[61911]: DEBUG oslo_concurrency.lockutils [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.679882] env[61911]: DEBUG nova.network.neutron [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Refreshing network info cache for port f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.701117] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5c21e90592b461f610ecb9a81f2033b7',container_format='bare',created_at=2024-10-10T16:00:09Z,direct_url=,disk_format='vmdk',id=d8718bf5-976e-48a0-b72e-8fea0db5e230,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1222072450-shelved',owner='e00325b35c074e92a5c80bf2a2df08dd',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-10T16:00:26Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 940.701884] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.701884] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 940.702081] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.702315] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 940.702589] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 940.702900] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 940.703164] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 940.703457] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 940.703725] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 940.704009] env[61911]: DEBUG nova.virt.hardware [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 940.704942] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fb07d6-df9b-4b20-beaf-a290ab7e5219 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.714017] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239b02c8-e2e9-4aef-9c30-9ac7d57527ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.728822] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:e6:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6d1cfde-6743-4e5f-ac24-8b71f63c293c', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.736688] env[61911]: DEBUG oslo.service.loopingcall [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.736997] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.737288] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06c31573-f292-427f-b813-b1d2b2c8486e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.757608] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.757608] env[61911]: value = "task-1251362" [ 940.757608] env[61911]: _type = "Task" [ 940.757608] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.765946] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251362, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.819991] env[61911]: DEBUG nova.compute.manager [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 940.820270] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 940.821213] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9376643-a525-40c3-ab17-d1556a9f0cec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.831793] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.832100] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37652e75-1bbc-4445-b5b6-1cdd65632008 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.837995] env[61911]: DEBUG oslo_vmware.api [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 940.837995] env[61911]: value = "task-1251363" [ 940.837995] env[61911]: _type = "Task" [ 940.837995] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.846167] env[61911]: DEBUG oslo_vmware.api [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251363, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.985494] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251361, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437868} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.985773] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 3c4c8943-5324-4c3e-b7e0-7ea5972bc026/3c4c8943-5324-4c3e-b7e0-7ea5972bc026.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.985999] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.986274] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d5e2ea2-61d3-4cd0-b0f1-ed9a7f0a06ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.996168] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 940.996168] env[61911]: value = "task-1251364" [ 940.996168] env[61911]: _type = "Task" [ 940.996168] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.004784] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251364, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.091782] env[61911]: DEBUG nova.scheduler.client.report [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 941.269359] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251362, 'name': CreateVM_Task, 'duration_secs': 0.424651} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.269544] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.270210] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.270383] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.270754] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.271358] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1537b858-5849-4f72-9401-3a05328928f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.275590] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 941.275590] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de03c0-edc2-80cf-1ccc-f9ba5eaf780a" [ 941.275590] env[61911]: _type = "Task" [ 941.275590] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.282583] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de03c0-edc2-80cf-1ccc-f9ba5eaf780a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.348891] env[61911]: DEBUG oslo_vmware.api [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251363, 'name': PowerOffVM_Task, 'duration_secs': 0.206814} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.349199] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.349351] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.349604] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd1147ac-3003-4962-86ff-a6ec6efbed96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.374083] env[61911]: DEBUG nova.network.neutron [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updated VIF entry in instance network info cache for port f6d1cfde-6743-4e5f-ac24-8b71f63c293c. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.374426] env[61911]: DEBUG nova.network.neutron [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.409536] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.409798] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.409969] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleting the datastore file [datastore1] 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.410297] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bae170e0-e900-4c56-b774-575849b404cd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.416999] env[61911]: DEBUG oslo_vmware.api [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 941.416999] env[61911]: value = "task-1251366" [ 941.416999] env[61911]: _type = "Task" [ 941.416999] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.424883] env[61911]: DEBUG oslo_vmware.api [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.505764] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251364, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093273} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.506055] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.506820] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d260c9c-42c8-42dc-bc14-dbb7d923938d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.528737] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 3c4c8943-5324-4c3e-b7e0-7ea5972bc026/3c4c8943-5324-4c3e-b7e0-7ea5972bc026.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.529023] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a875c28f-0624-4c4e-b0ee-c5dec3aacafd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.549311] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 941.549311] env[61911]: value = "task-1251367" [ 941.549311] env[61911]: _type = "Task" [ 941.549311] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.558759] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.596871] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.597430] env[61911]: DEBUG nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 941.601041] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.729s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.601190] env[61911]: DEBUG nova.objects.instance [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lazy-loading 'resources' on Instance uuid 0ebf8d34-54c0-453e-b2e2-820feed6ee61 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.786282] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.786559] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Processing image d8718bf5-976e-48a0-b72e-8fea0db5e230 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.786806] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230/d8718bf5-976e-48a0-b72e-8fea0db5e230.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.786959] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230/d8718bf5-976e-48a0-b72e-8fea0db5e230.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.787168] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.787430] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8fc98bf-b42a-4623-af65-f6879e0eeeb9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.796042] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.796236] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.796952] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cc60d82-e1c9-44ca-983d-ccb5525a22b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.801801] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 941.801801] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c88341-6cc1-ac29-0ea4-6f14b54fedc3" [ 941.801801] env[61911]: _type = "Task" [ 941.801801] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.809301] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c88341-6cc1-ac29-0ea4-6f14b54fedc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.877296] env[61911]: DEBUG oslo_concurrency.lockutils [req-13de1542-a2c2-4573-acbf-e8379a89235f req-31a2e4af-3929-4de1-94b3-189a8037b483 service nova] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.881141] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.881468] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a685b9b-1744-44ac-9598-9fb8ce9d6944 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.888725] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 941.888725] env[61911]: value = "task-1251368" [ 941.888725] env[61911]: _type = "Task" [ 941.888725] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.896835] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251368, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.926274] env[61911]: DEBUG oslo_vmware.api [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153026} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.926532] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.926715] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.926897] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.927092] env[61911]: INFO nova.compute.manager [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 941.927334] env[61911]: DEBUG oslo.service.loopingcall [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.927527] env[61911]: DEBUG nova.compute.manager [-] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 941.927621] env[61911]: DEBUG nova.network.neutron [-] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.059144] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251367, 'name': ReconfigVM_Task, 'duration_secs': 0.26923} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.059433] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 3c4c8943-5324-4c3e-b7e0-7ea5972bc026/3c4c8943-5324-4c3e-b7e0-7ea5972bc026.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.060115] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51908dc8-9ffe-4ad4-bcb3-29eea80b6654 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.066429] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 942.066429] env[61911]: value = "task-1251369" [ 942.066429] env[61911]: _type = "Task" [ 942.066429] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.079467] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251369, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.103105] env[61911]: DEBUG nova.compute.utils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.107681] env[61911]: DEBUG nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 942.107891] env[61911]: DEBUG nova.network.neutron [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 942.150665] env[61911]: DEBUG nova.policy [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b034026eede4f5ca4757b552f7dca51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5be8ec1d59a14d618906ec32d0e134e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.314010] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Preparing fetch location {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 942.314293] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Fetch image to [datastore1] OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33/OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33.vmdk {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 942.314483] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Downloading stream optimized image d8718bf5-976e-48a0-b72e-8fea0db5e230 to [datastore1] OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33/OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33.vmdk on the data store datastore1 as vApp {{(pid=61911) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 942.314654] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Downloading image file data d8718bf5-976e-48a0-b72e-8fea0db5e230 to the ESX as VM named 'OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33' {{(pid=61911) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 942.316944] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78271ff7-c303-4ead-bbf1-23fa7a460dc1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.324238] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369a622c-f21d-4827-ad9d-71e84146f29d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.357896] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0488ccc0-40aa-4b25-89e5-22ace7c3ed25 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.365705] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d62b23-c274-49aa-b883-3dbf3461d168 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.399698] env[61911]: DEBUG nova.compute.provider_tree [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.401422] env[61911]: DEBUG nova.network.neutron [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Successfully created port: 5937b7a1-0418-4e62-9605-d52f38ca02b4 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 942.409033] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251368, 'name': PowerOffVM_Task, 'duration_secs': 0.182105} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.409822] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.410640] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b383a6e-5b19-4966-96bc-db598b8f833d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.432459] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 942.432459] env[61911]: value = "resgroup-9" [ 942.432459] env[61911]: _type = "ResourcePool" [ 942.432459] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 942.433232] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc1f9e0-5aec-4427-880b-83c2c7836e7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.435865] env[61911]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-e9dcf6d8-e031-411d-ad04-8aacdefc92a3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.457914] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lease: (returnval){ [ 942.457914] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268084b-7915-eb16-a2d2-1017fee75a1f" [ 942.457914] env[61911]: _type = "HttpNfcLease" [ 942.457914] env[61911]: } obtained for vApp import into resource pool (val){ [ 942.457914] env[61911]: value = "resgroup-9" [ 942.457914] env[61911]: _type = "ResourcePool" [ 942.457914] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 942.458188] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the lease: (returnval){ [ 942.458188] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268084b-7915-eb16-a2d2-1017fee75a1f" [ 942.458188] env[61911]: _type = "HttpNfcLease" [ 942.458188] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 942.464070] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 942.464070] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268084b-7915-eb16-a2d2-1017fee75a1f" [ 942.464070] env[61911]: _type = "HttpNfcLease" [ 942.464070] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 942.474600] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.474856] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-390251c9-d1a9-4f6e-a07a-125ea34af10a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.480297] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 942.480297] env[61911]: value = "task-1251371" [ 942.480297] env[61911]: _type = "Task" [ 942.480297] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.492768] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251371, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.576719] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251369, 'name': Rename_Task, 'duration_secs': 0.144899} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.577303] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.577303] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec5b9a81-53b4-4df8-88dc-71fcf534a7db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.583641] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 942.583641] env[61911]: value = "task-1251372" [ 942.583641] env[61911]: _type = "Task" [ 942.583641] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.591673] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251372, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.614020] env[61911]: DEBUG nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 942.716978] env[61911]: DEBUG nova.compute.manager [req-36530372-67af-4005-b836-67fc4034f353 req-94879030-43a8-42bd-89d2-88b4a11714b4 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Received event network-vif-deleted-9cdfed52-77e2-4eb9-9aac-f603cdea8d01 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 942.717212] env[61911]: INFO nova.compute.manager [req-36530372-67af-4005-b836-67fc4034f353 req-94879030-43a8-42bd-89d2-88b4a11714b4 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Neutron deleted interface 9cdfed52-77e2-4eb9-9aac-f603cdea8d01; detaching it from the instance and deleting it from the info cache [ 942.717393] env[61911]: DEBUG nova.network.neutron [req-36530372-67af-4005-b836-67fc4034f353 req-94879030-43a8-42bd-89d2-88b4a11714b4 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.787163] env[61911]: DEBUG nova.network.neutron [-] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.906152] env[61911]: DEBUG nova.scheduler.client.report [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 942.967332] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 942.967332] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268084b-7915-eb16-a2d2-1017fee75a1f" [ 942.967332] env[61911]: _type = "HttpNfcLease" [ 942.967332] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 942.969248] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 942.969248] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5268084b-7915-eb16-a2d2-1017fee75a1f" [ 942.969248] env[61911]: _type = "HttpNfcLease" [ 942.969248] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 942.969248] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd3f25e-ad4f-4041-af9c-230f7cb0c3b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.976496] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ad6be0-7198-cbcc-7a55-a1fe3256cb3a/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 942.976691] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ad6be0-7198-cbcc-7a55-a1fe3256cb3a/disk-0.vmdk. {{(pid=61911) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 943.045250] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cbed76bd-252c-48cf-86a9-72aba4417882 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.047981] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 943.048257] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.048528] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.048677] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.048858] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.049539] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f41cf2d-e97b-43ef-9cb5-0527a52e7c4f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.058897] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.059089] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.059857] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1586fbc5-90c6-4489-91a5-ef6552e147dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.065874] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 943.065874] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb587a-8d2e-7db1-3355-096f14485a80" [ 943.065874] env[61911]: _type = "Task" [ 943.065874] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.074994] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb587a-8d2e-7db1-3355-096f14485a80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.093050] env[61911]: DEBUG oslo_vmware.api [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251372, 'name': PowerOnVM_Task, 'duration_secs': 0.475536} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.093353] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.093568] env[61911]: INFO nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Took 6.83 seconds to spawn the instance on the hypervisor. [ 943.093783] env[61911]: DEBUG nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 943.094598] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031b58c8-1bde-4a37-8638-04bfd78449e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.220598] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d6f9c83-9849-4486-bfff-b7eade5c19bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.231646] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe5a89b-7354-4211-ba15-026665481179 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.260570] env[61911]: DEBUG nova.compute.manager [req-36530372-67af-4005-b836-67fc4034f353 req-94879030-43a8-42bd-89d2-88b4a11714b4 service nova] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Detach interface failed, port_id=9cdfed52-77e2-4eb9-9aac-f603cdea8d01, reason: Instance 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 943.291801] env[61911]: INFO nova.compute.manager [-] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Took 1.36 seconds to deallocate network for instance. [ 943.411547] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.415329] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.153s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.415593] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.419209] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.768s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.420811] env[61911]: INFO nova.compute.claims [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.434863] env[61911]: INFO nova.scheduler.client.report [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Deleted allocations for instance 0ebf8d34-54c0-453e-b2e2-820feed6ee61 [ 943.442923] env[61911]: INFO nova.scheduler.client.report [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted allocations for instance a35a11f1-3d4b-439e-9517-fa5a9f43af60 [ 943.581046] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb587a-8d2e-7db1-3355-096f14485a80, 'name': SearchDatastore_Task, 'duration_secs': 0.011339} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.583300] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0307c894-a08a-4084-b550-4fcf52740f78 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.589635] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 943.589635] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52dd3636-4ecc-63de-ed1f-469bad43afbb" [ 943.589635] env[61911]: _type = "Task" [ 943.589635] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.599434] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52dd3636-4ecc-63de-ed1f-469bad43afbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.613079] env[61911]: INFO nova.compute.manager [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Took 19.68 seconds to build instance. [ 943.620894] env[61911]: DEBUG nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 943.648597] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 943.648882] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.649069] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 943.649270] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.649423] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 943.649575] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 943.649789] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 943.649952] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 943.650237] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 943.650463] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 943.650660] env[61911]: DEBUG nova.virt.hardware [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 943.651854] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e92172d-3c45-4b77-a000-015472df2e6d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.664433] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47eeb1c-1417-45e1-a9be-f01320476a5b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.798709] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.897121] env[61911]: DEBUG nova.compute.manager [req-8f2e8bb0-1c7c-42ee-a89a-4213ab44efb9 req-99ca1a23-12e9-499d-a34d-ac8dedf6720f service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Received event network-vif-plugged-5937b7a1-0418-4e62-9605-d52f38ca02b4 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 943.897121] env[61911]: DEBUG oslo_concurrency.lockutils [req-8f2e8bb0-1c7c-42ee-a89a-4213ab44efb9 req-99ca1a23-12e9-499d-a34d-ac8dedf6720f service nova] Acquiring lock "f386424d-953d-431d-a56b-8542f2a3458c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.897121] env[61911]: DEBUG oslo_concurrency.lockutils [req-8f2e8bb0-1c7c-42ee-a89a-4213ab44efb9 req-99ca1a23-12e9-499d-a34d-ac8dedf6720f service nova] Lock "f386424d-953d-431d-a56b-8542f2a3458c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.897121] env[61911]: DEBUG oslo_concurrency.lockutils [req-8f2e8bb0-1c7c-42ee-a89a-4213ab44efb9 req-99ca1a23-12e9-499d-a34d-ac8dedf6720f service nova] Lock "f386424d-953d-431d-a56b-8542f2a3458c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.897121] env[61911]: DEBUG nova.compute.manager [req-8f2e8bb0-1c7c-42ee-a89a-4213ab44efb9 req-99ca1a23-12e9-499d-a34d-ac8dedf6720f service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] No waiting events found dispatching network-vif-plugged-5937b7a1-0418-4e62-9605-d52f38ca02b4 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 943.897121] env[61911]: WARNING nova.compute.manager [req-8f2e8bb0-1c7c-42ee-a89a-4213ab44efb9 req-99ca1a23-12e9-499d-a34d-ac8dedf6720f service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Received unexpected event network-vif-plugged-5937b7a1-0418-4e62-9605-d52f38ca02b4 for instance with vm_state building and task_state spawning. [ 943.943267] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0c212c3c-b3f0-4352-9df3-7d2b12d1198c tempest-ImagesNegativeTestJSON-525842046 tempest-ImagesNegativeTestJSON-525842046-project-member] Lock "0ebf8d34-54c0-453e-b2e2-820feed6ee61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.824s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.952322] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b9e21d61-1495-4667-bc6a-0dc5fc3d2764 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "a35a11f1-3d4b-439e-9517-fa5a9f43af60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.198s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.976940] env[61911]: DEBUG nova.network.neutron [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Successfully updated port: 5937b7a1-0418-4e62-9605-d52f38ca02b4 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.102035] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52dd3636-4ecc-63de-ed1f-469bad43afbb, 'name': SearchDatastore_Task, 'duration_secs': 0.013262} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.105342] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.105685] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. {{(pid=61911) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 944.106044] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d785b9d5-87e3-4c10-b6ea-641d4d6eecc9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.114892] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1c21abce-6dd0-4561-beab-782ffe64e3f7 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.190s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.115242] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 944.115242] env[61911]: value = "task-1251373" [ 944.115242] env[61911]: _type = "Task" [ 944.115242] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.128668] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251373, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.340604] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Completed reading data from the image iterator. {{(pid=61911) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 944.340966] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ad6be0-7198-cbcc-7a55-a1fe3256cb3a/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 944.344295] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca2e6e5-0284-42f6-8949-af218e142cc4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.352064] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ad6be0-7198-cbcc-7a55-a1fe3256cb3a/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 944.352359] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ad6be0-7198-cbcc-7a55-a1fe3256cb3a/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 944.352714] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4becda9f-eb96-4a8d-9606-57ed6ccb2f4d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.479403] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.479790] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.480059] env[61911]: DEBUG nova.network.neutron [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.629115] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251373, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.643044] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c0e4d6-c743-4387-adce-0ca81dd14458 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.647287] env[61911]: DEBUG oslo_vmware.rw_handles [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ad6be0-7198-cbcc-7a55-a1fe3256cb3a/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 944.647531] env[61911]: INFO nova.virt.vmwareapi.images [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Downloaded image file data d8718bf5-976e-48a0-b72e-8fea0db5e230 [ 944.648786] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75afdfb-3cc8-4354-8ca8-965493accca5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.654935] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeeee5e8-e4fe-4c3e-95b8-07e0533e2d52 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.670151] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10c07048-4db4-49a5-87c5-a0b45f040fea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.701225] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfc12dd-c640-4591-8192-7080de69fbf5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.709894] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f778f2-4bd6-4169-998c-1974a65742b5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.730173] env[61911]: DEBUG nova.compute.provider_tree [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.734431] env[61911]: INFO nova.virt.vmwareapi.images [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] The imported VM was unregistered [ 944.735825] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Caching image {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 944.736119] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Creating directory with path [datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.736346] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bea0e2f-d287-46ad-b0ae-6dc1d2ce925c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.761302] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Created directory with path [datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.761534] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33/OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33.vmdk to [datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230/d8718bf5-976e-48a0-b72e-8fea0db5e230.vmdk. {{(pid=61911) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 944.761875] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-3967cfb5-5613-4489-a394-4c80facc59ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.776586] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 944.776586] env[61911]: value = "task-1251375" [ 944.776586] env[61911]: _type = "Task" [ 944.776586] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.788070] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251375, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.041387] env[61911]: DEBUG nova.network.neutron [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.130345] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251373, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595368} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.130661] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk. [ 945.131567] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046ed7bf-98f5-46a2-8bdd-71387023a5ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.163744] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.164347] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f8f106b-2534-4a62-9a7f-f49ae82274dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.189334] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 945.189334] env[61911]: value = "task-1251376" [ 945.189334] env[61911]: _type = "Task" [ 945.189334] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.198586] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251376, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.234914] env[61911]: DEBUG nova.scheduler.client.report [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 945.287307] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251375, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.323491] env[61911]: DEBUG nova.network.neutron [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance_info_cache with network_info: [{"id": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "address": "fa:16:3e:ef:fb:2e", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5937b7a1-04", "ovs_interfaceid": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.344508] env[61911]: DEBUG nova.compute.manager [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 945.344508] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c43c50-158f-4197-8e8b-4b30e8f87374 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.700450] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.739908] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.740643] env[61911]: DEBUG nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 945.743496] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.072s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.745114] env[61911]: INFO nova.compute.claims [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.789715] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251375, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.828848] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.828848] env[61911]: DEBUG nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Instance network_info: |[{"id": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "address": "fa:16:3e:ef:fb:2e", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5937b7a1-04", "ovs_interfaceid": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 945.828848] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:fb:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5937b7a1-0418-4e62-9605-d52f38ca02b4', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.836479] env[61911]: DEBUG oslo.service.loopingcall [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.837171] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.837440] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-90569bba-512a-4730-a06f-3774be45a7c9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.857441] env[61911]: INFO nova.compute.manager [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] instance snapshotting [ 945.863363] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b6549c-77cc-4e2c-9d7b-370808fe79c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.867208] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.867208] env[61911]: value = "task-1251377" [ 945.867208] env[61911]: _type = "Task" [ 945.867208] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.888641] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82ea798-5265-4f49-b961-ff8f7e0b962b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.894794] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251377, 'name': CreateVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.960334] env[61911]: DEBUG nova.compute.manager [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Received event network-changed-5937b7a1-0418-4e62-9605-d52f38ca02b4 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 945.960597] env[61911]: DEBUG nova.compute.manager [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Refreshing instance network info cache due to event network-changed-5937b7a1-0418-4e62-9605-d52f38ca02b4. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 945.960881] env[61911]: DEBUG oslo_concurrency.lockutils [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] Acquiring lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.961010] env[61911]: DEBUG oslo_concurrency.lockutils [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] Acquired lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.961175] env[61911]: DEBUG nova.network.neutron [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Refreshing network info cache for port 5937b7a1-0418-4e62-9605-d52f38ca02b4 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.201305] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.252023] env[61911]: DEBUG nova.compute.utils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 946.255798] env[61911]: DEBUG nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 946.260023] env[61911]: DEBUG nova.network.neutron [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 946.287889] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251375, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.312220] env[61911]: DEBUG nova.policy [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a5252f9309f4204a3a3fd7d865db6e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b285793da304b5e8f9736098ef163c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 946.380797] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251377, 'name': CreateVM_Task, 'duration_secs': 0.465676} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.381174] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.382331] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.382422] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.382850] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.383166] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cab8c336-c178-4aeb-a663-2dccabb4e573 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.391471] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 946.391471] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529519d4-8c2d-2978-7ab1-f8866eea37fd" [ 946.391471] env[61911]: _type = "Task" [ 946.391471] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.405116] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529519d4-8c2d-2978-7ab1-f8866eea37fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.407529] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 946.407938] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1f55a22b-cd47-4fd8-adab-ff3a25c2bb76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.417465] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 946.417465] env[61911]: value = "task-1251378" [ 946.417465] env[61911]: _type = "Task" [ 946.417465] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.427916] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251378, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.667551] env[61911]: DEBUG nova.network.neutron [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Successfully created port: def5852e-6e08-47c9-9179-88d9b4161e8d {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.702023] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.756527] env[61911]: DEBUG nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 946.792069] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251375, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.794261] env[61911]: DEBUG nova.network.neutron [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updated VIF entry in instance network info cache for port 5937b7a1-0418-4e62-9605-d52f38ca02b4. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 946.794596] env[61911]: DEBUG nova.network.neutron [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance_info_cache with network_info: [{"id": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "address": "fa:16:3e:ef:fb:2e", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5937b7a1-04", "ovs_interfaceid": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.904853] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529519d4-8c2d-2978-7ab1-f8866eea37fd, 'name': SearchDatastore_Task, 'duration_secs': 0.096199} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.908081] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.908578] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.908731] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.908892] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.909118] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.909770] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54f6c559-7c86-4283-a4b3-0c7756d647fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.931464] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.931687] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.932551] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251378, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.932761] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7651082-7265-43a2-804d-e1b04a6d3ca1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.939604] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 946.939604] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e9d56e-2f4b-a808-348e-c6cb669be927" [ 946.939604] env[61911]: _type = "Task" [ 946.939604] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.949218] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e9d56e-2f4b-a808-348e-c6cb669be927, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.018246] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff0456a-b3ae-4ac3-8e08-0376a761ff97 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.028270] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9403a9e2-2371-47ac-aec8-354a0d8204e7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.073237] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b936c882-5443-449a-884a-39fe8b733aa4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.084436] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f33ab60-c468-404f-9aac-291c7ae79266 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.104452] env[61911]: DEBUG nova.compute.provider_tree [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.203341] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.288676] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251375, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.300729] env[61911]: DEBUG oslo_concurrency.lockutils [req-8b5cf522-883a-4abd-9227-7911f9313417 req-3b4581b1-eae0-4b74-9f72-2c91000c4998 service nova] Releasing lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.428735] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251378, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.449882] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e9d56e-2f4b-a808-348e-c6cb669be927, 'name': SearchDatastore_Task, 'duration_secs': 0.090682} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.450755] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b271338-306a-4858-8758-f11b13a9660d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.456168] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 947.456168] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52da8d73-165d-a336-3ad5-c1458b902010" [ 947.456168] env[61911]: _type = "Task" [ 947.456168] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.464229] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52da8d73-165d-a336-3ad5-c1458b902010, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.608497] env[61911]: DEBUG nova.scheduler.client.report [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 947.703892] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251376, 'name': ReconfigVM_Task, 'duration_secs': 2.320966} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.705033] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c/0dfe2ff1-43fd-4529-93f1-daaccc0711cf-rescue.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.705152] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1e1b78-2c80-4674-825d-0930eeabc259 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.733286] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd527759-2205-46f8-bdd5-c1d52ee8b5d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.748312] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 947.748312] env[61911]: value = "task-1251379" [ 947.748312] env[61911]: _type = "Task" [ 947.748312] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.756812] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251379, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.766288] env[61911]: DEBUG nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 947.788455] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251375, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.603247} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.791384] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33/OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33.vmdk to [datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230/d8718bf5-976e-48a0-b72e-8fea0db5e230.vmdk. [ 947.791384] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Cleaning up location [datastore1] OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 947.791384] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ff376f53-4fd1-4dc5-8fd0-03df12b83a33 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.791665] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96b8e4a1-dea0-45f1-936a-a1cb6ec5817b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.797666] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 947.797953] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.798124] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 947.798311] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.798467] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 947.798620] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 947.799146] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 947.799403] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 947.799672] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 947.799906] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 947.800289] env[61911]: DEBUG nova.virt.hardware [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 947.801179] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836ba85a-bb21-4fc8-a3ee-66d4066427ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.805087] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 947.805087] env[61911]: value = "task-1251380" [ 947.805087] env[61911]: _type = "Task" [ 947.805087] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.812372] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11402892-4a06-4925-9c9b-d4af6edb6d1c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.822823] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251380, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.929426] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251378, 'name': CreateSnapshot_Task, 'duration_secs': 1.155402} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.929724] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 947.930655] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78bd2df1-076d-4128-9038-4978cc8fee02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.966337] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52da8d73-165d-a336-3ad5-c1458b902010, 'name': SearchDatastore_Task, 'duration_secs': 0.020265} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.966665] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.967147] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] f386424d-953d-431d-a56b-8542f2a3458c/f386424d-953d-431d-a56b-8542f2a3458c.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.967297] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34379181-8530-4f44-b635-9973064a6904 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.974076] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 947.974076] env[61911]: value = "task-1251381" [ 947.974076] env[61911]: _type = "Task" [ 947.974076] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.984764] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.118957] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.118957] env[61911]: DEBUG nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 948.121222] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.823s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.121532] env[61911]: DEBUG nova.objects.instance [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lazy-loading 'resources' on Instance uuid 66472f43-537d-4eb3-8d49-d40627a8809d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.259493] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251379, 'name': ReconfigVM_Task, 'duration_secs': 0.217932} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.259907] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.260101] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c166844-3359-4b6e-adca-e7d373a549ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.268039] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 948.268039] env[61911]: value = "task-1251382" [ 948.268039] env[61911]: _type = "Task" [ 948.268039] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.277036] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251382, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.315632] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251380, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.043043} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.315786] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.315959] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230/d8718bf5-976e-48a0-b72e-8fea0db5e230.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.316229] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230/d8718bf5-976e-48a0-b72e-8fea0db5e230.vmdk to [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.316507] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c7f0175-9ac0-42ae-a0e2-9adee49dc849 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.323963] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 948.323963] env[61911]: value = "task-1251383" [ 948.323963] env[61911]: _type = "Task" [ 948.323963] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.332954] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251383, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.406439] env[61911]: DEBUG nova.network.neutron [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Successfully updated port: def5852e-6e08-47c9-9179-88d9b4161e8d {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.453084] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 948.454342] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-505800c2-6375-4098-962f-f15b92376ea3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.463252] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 948.463252] env[61911]: value = "task-1251384" [ 948.463252] env[61911]: _type = "Task" [ 948.463252] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.473227] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251384, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.482789] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471659} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.483140] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] f386424d-953d-431d-a56b-8542f2a3458c/f386424d-953d-431d-a56b-8542f2a3458c.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.483574] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.483856] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-040d3b27-adae-442f-83f2-46420dd138f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.493159] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 948.493159] env[61911]: value = "task-1251385" [ 948.493159] env[61911]: _type = "Task" [ 948.493159] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.501618] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251385, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.626243] env[61911]: DEBUG nova.compute.utils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.643080] env[61911]: DEBUG nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 948.643080] env[61911]: DEBUG nova.network.neutron [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.716776] env[61911]: DEBUG nova.policy [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62586445653844078feeeb731bd16f62', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e6ca33655884f16b4c0a6fcdb31ec12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.761129] env[61911]: DEBUG nova.compute.manager [req-fe531354-3c51-46a3-9f7f-cfcc5d886827 req-074f33f2-4508-441b-ade6-884acb210220 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Received event network-vif-plugged-def5852e-6e08-47c9-9179-88d9b4161e8d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 948.761129] env[61911]: DEBUG oslo_concurrency.lockutils [req-fe531354-3c51-46a3-9f7f-cfcc5d886827 req-074f33f2-4508-441b-ade6-884acb210220 service nova] Acquiring lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.761129] env[61911]: DEBUG oslo_concurrency.lockutils [req-fe531354-3c51-46a3-9f7f-cfcc5d886827 req-074f33f2-4508-441b-ade6-884acb210220 service nova] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.761609] env[61911]: DEBUG oslo_concurrency.lockutils [req-fe531354-3c51-46a3-9f7f-cfcc5d886827 req-074f33f2-4508-441b-ade6-884acb210220 service nova] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.761647] env[61911]: DEBUG nova.compute.manager [req-fe531354-3c51-46a3-9f7f-cfcc5d886827 req-074f33f2-4508-441b-ade6-884acb210220 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] No waiting events found dispatching network-vif-plugged-def5852e-6e08-47c9-9179-88d9b4161e8d {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 948.762557] env[61911]: WARNING nova.compute.manager [req-fe531354-3c51-46a3-9f7f-cfcc5d886827 req-074f33f2-4508-441b-ade6-884acb210220 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Received unexpected event network-vif-plugged-def5852e-6e08-47c9-9179-88d9b4161e8d for instance with vm_state building and task_state spawning. [ 948.785353] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251382, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.833753] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251383, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.890591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1f907b-d9e7-4d86-91b7-7e6e7f6c4f69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.898974] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054089bf-b261-406f-ab4c-8394ac029040 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.942907] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-ec53a8ee-df42-4003-a933-17aa4c90c3aa" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.943084] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-ec53a8ee-df42-4003-a933-17aa4c90c3aa" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.943247] env[61911]: DEBUG nova.network.neutron [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.948131] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f46a99-e4a0-425d-afe7-d2ce1d0a5605 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.957521] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2a09f8-12ec-4c88-aba4-93cf66082ea8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.977945] env[61911]: DEBUG nova.compute.provider_tree [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.988631] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251384, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.004089] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251385, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072282} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.004850] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.005236] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20735fcf-f82b-4c6d-b8f3-cd48c078b2e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.029372] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] f386424d-953d-431d-a56b-8542f2a3458c/f386424d-953d-431d-a56b-8542f2a3458c.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.029755] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06156c27-44f9-4d3c-a343-22693e7637d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.051659] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 949.051659] env[61911]: value = "task-1251386" [ 949.051659] env[61911]: _type = "Task" [ 949.051659] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.060498] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251386, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.122209] env[61911]: DEBUG nova.network.neutron [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Successfully created port: 33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.138611] env[61911]: DEBUG nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 949.282509] env[61911]: DEBUG oslo_vmware.api [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251382, 'name': PowerOnVM_Task, 'duration_secs': 0.877363} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.282509] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.285323] env[61911]: DEBUG nova.compute.manager [None req-fd5b0dc0-fc4a-413f-8904-ac882a4d1fa1 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 949.286460] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720dcdcc-004e-43df-9042-497500b5aac2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.337053] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251383, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.476310] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251384, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.481482] env[61911]: DEBUG nova.scheduler.client.report [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 949.496258] env[61911]: DEBUG nova.network.neutron [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 949.567209] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251386, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.760386] env[61911]: DEBUG nova.network.neutron [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Updating instance_info_cache with network_info: [{"id": "def5852e-6e08-47c9-9179-88d9b4161e8d", "address": "fa:16:3e:6e:0b:7b", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdef5852e-6e", "ovs_interfaceid": "def5852e-6e08-47c9-9179-88d9b4161e8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.842878] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251383, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.975490] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251384, 'name': CloneVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.999174] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.003151] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.203s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.004057] env[61911]: DEBUG nova.objects.instance [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lazy-loading 'resources' on Instance uuid 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.024294] env[61911]: INFO nova.scheduler.client.report [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Deleted allocations for instance 66472f43-537d-4eb3-8d49-d40627a8809d [ 950.065470] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251386, 'name': ReconfigVM_Task, 'duration_secs': 0.662695} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.065800] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Reconfigured VM instance instance-00000055 to attach disk [datastore1] f386424d-953d-431d-a56b-8542f2a3458c/f386424d-953d-431d-a56b-8542f2a3458c.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.066500] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-817b2489-c2a8-4561-9dbb-0f91232c91f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.073090] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 950.073090] env[61911]: value = "task-1251387" [ 950.073090] env[61911]: _type = "Task" [ 950.073090] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.089341] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251387, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.154757] env[61911]: DEBUG nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 950.186705] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 950.186811] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.186956] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 950.188437] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.188437] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 950.188437] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 950.188437] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 950.188437] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 950.188437] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 950.188437] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 950.188812] env[61911]: DEBUG nova.virt.hardware [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 950.189475] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd5e075-e668-45b0-8f00-10839b7d9928 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.203206] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57131bee-666e-40d6-a8fe-a274cb7f9b2e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.264096] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-ec53a8ee-df42-4003-a933-17aa4c90c3aa" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.264586] env[61911]: DEBUG nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Instance network_info: |[{"id": "def5852e-6e08-47c9-9179-88d9b4161e8d", "address": "fa:16:3e:6e:0b:7b", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdef5852e-6e", "ovs_interfaceid": "def5852e-6e08-47c9-9179-88d9b4161e8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 950.264949] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:0b:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'def5852e-6e08-47c9-9179-88d9b4161e8d', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.274833] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating folder: Project (1b285793da304b5e8f9736098ef163c4). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 950.274833] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8072f52-959d-42b7-8b98-9d012560a474 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.287174] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created folder: Project (1b285793da304b5e8f9736098ef163c4) in parent group-v269521. [ 950.287174] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating folder: Instances. Parent ref: group-v269639. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 950.287174] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26c94b54-7e98-4613-8e07-2d29b4ef0188 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.299135] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created folder: Instances in parent group-v269639. [ 950.299689] env[61911]: DEBUG oslo.service.loopingcall [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.299689] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 950.300054] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f7121d4-f532-4939-9639-0711c24c8c84 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.320815] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.320815] env[61911]: value = "task-1251390" [ 950.320815] env[61911]: _type = "Task" [ 950.320815] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.335612] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251390, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.343966] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251383, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.475725] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251384, 'name': CloneVM_Task, 'duration_secs': 1.517548} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.476071] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Created linked-clone VM from snapshot [ 950.476935] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ba8dea-5e16-45cc-ae22-4928b93918e7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.485816] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Uploading image 0127f673-151a-45e0-aedf-e4b30b4dcf9d {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 950.506701] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 950.511322] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e94220c6-7944-45df-9e6d-9d99d569b072 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.523512] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 950.523512] env[61911]: value = "task-1251391" [ 950.523512] env[61911]: _type = "Task" [ 950.523512] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.535654] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251391, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.535654] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ff8cec79-026d-4de1-81fe-16baf7d6716e tempest-ServersAdminTestJSON-1731360060 tempest-ServersAdminTestJSON-1731360060-project-member] Lock "66472f43-537d-4eb3-8d49-d40627a8809d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.685s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.588378] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251387, 'name': Rename_Task, 'duration_secs': 0.235764} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.589495] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.589877] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf1e32d7-40b9-41bf-a6e0-79ca2ba240ed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.600049] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 950.600049] env[61911]: value = "task-1251392" [ 950.600049] env[61911]: _type = "Task" [ 950.600049] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.615932] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251392, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.659922] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.660811] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.661096] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 950.746113] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edb2fa5-2ba2-4ea9-8c00-719bc1f5af31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.755844] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36b8561-8f9a-4ea3-aca3-4f498c452172 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.793374] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1680d2-b487-4762-b89d-08a7d2d8f057 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.802793] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ffa2a4-ec51-45d4-acd4-969417603085 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.813119] env[61911]: DEBUG nova.compute.manager [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Received event network-changed-def5852e-6e08-47c9-9179-88d9b4161e8d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 950.813411] env[61911]: DEBUG nova.compute.manager [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Refreshing instance network info cache due to event network-changed-def5852e-6e08-47c9-9179-88d9b4161e8d. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 950.813762] env[61911]: DEBUG oslo_concurrency.lockutils [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] Acquiring lock "refresh_cache-ec53a8ee-df42-4003-a933-17aa4c90c3aa" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.814095] env[61911]: DEBUG oslo_concurrency.lockutils [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] Acquired lock "refresh_cache-ec53a8ee-df42-4003-a933-17aa4c90c3aa" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.814379] env[61911]: DEBUG nova.network.neutron [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Refreshing network info cache for port def5852e-6e08-47c9-9179-88d9b4161e8d {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.826487] env[61911]: DEBUG nova.compute.provider_tree [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.848126] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251390, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.851680] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251383, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.035026] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251391, 'name': Destroy_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.079081] env[61911]: DEBUG nova.network.neutron [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Successfully updated port: 33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.113590] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251392, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.123607] env[61911]: DEBUG nova.network.neutron [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Updated VIF entry in instance network info cache for port def5852e-6e08-47c9-9179-88d9b4161e8d. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 951.123607] env[61911]: DEBUG nova.network.neutron [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Updating instance_info_cache with network_info: [{"id": "def5852e-6e08-47c9-9179-88d9b4161e8d", "address": "fa:16:3e:6e:0b:7b", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdef5852e-6e", "ovs_interfaceid": "def5852e-6e08-47c9-9179-88d9b4161e8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.161426] env[61911]: INFO nova.compute.manager [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Unrescuing [ 951.162751] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.162751] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquired lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.163578] env[61911]: DEBUG nova.network.neutron [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.331441] env[61911]: DEBUG nova.scheduler.client.report [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 951.343517] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251390, 'name': CreateVM_Task, 'duration_secs': 0.529041} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.345373] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 951.353421] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.353597] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.353942] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.354584] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.354802] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.356622] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251383, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.630136} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.357369] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8edfb0a8-f850-4ac4-a578-c591745711e7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.360430] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d8718bf5-976e-48a0-b72e-8fea0db5e230/d8718bf5-976e-48a0-b72e-8fea0db5e230.vmdk to [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 951.361591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c267e1cf-e7d4-48cc-be65-f415853b8e33 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.369816] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 951.369816] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523dee98-7381-34b2-b4ac-e19c5b95e7a0" [ 951.369816] env[61911]: _type = "Task" [ 951.369816] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.393027] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.396689] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8de6c09d-6794-489b-a3e4-a3e633deaa87 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.419200] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]523dee98-7381-34b2-b4ac-e19c5b95e7a0, 'name': SearchDatastore_Task, 'duration_secs': 0.014307} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.420941] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.421213] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.421471] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.421688] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.421958] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.422506] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 951.422506] env[61911]: value = "task-1251393" [ 951.422506] env[61911]: _type = "Task" [ 951.422506] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.422792] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34a14328-7092-42e7-bdbf-c6af41f4fd31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.435745] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.437191] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.437454] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 951.440226] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a5b0d62-fe86-4db2-8fa1-41f139316ed3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.444357] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 951.444357] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e6f9fe-f8aa-5ebe-5b7a-c96be4629c33" [ 951.444357] env[61911]: _type = "Task" [ 951.444357] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.452994] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e6f9fe-f8aa-5ebe-5b7a-c96be4629c33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.534270] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251391, 'name': Destroy_Task, 'duration_secs': 0.741458} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.534270] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Destroyed the VM [ 951.534270] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 951.534270] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2ead9523-6100-498b-abea-d5be0157264e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.541968] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 951.541968] env[61911]: value = "task-1251394" [ 951.541968] env[61911]: _type = "Task" [ 951.541968] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.550753] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251394, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.584998] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.585313] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.585625] env[61911]: DEBUG nova.network.neutron [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.612170] env[61911]: DEBUG oslo_vmware.api [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251392, 'name': PowerOnVM_Task, 'duration_secs': 0.88738} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.612451] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.612752] env[61911]: INFO nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Took 7.99 seconds to spawn the instance on the hypervisor. [ 951.613071] env[61911]: DEBUG nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 951.613888] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9adf2dd3-7a9f-41a5-b893-4c38a4c42acf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.625888] env[61911]: DEBUG oslo_concurrency.lockutils [req-a2f1d805-0f74-4b41-b942-f1fffadc5872 req-931f516c-0fe6-4ab7-bfa9-ca3cb64e1bc0 service nova] Releasing lock "refresh_cache-ec53a8ee-df42-4003-a933-17aa4c90c3aa" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.837068] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.858113] env[61911]: DEBUG nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 951.869515] env[61911]: INFO nova.scheduler.client.report [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted allocations for instance 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a [ 951.935564] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.954515] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e6f9fe-f8aa-5ebe-5b7a-c96be4629c33, 'name': SearchDatastore_Task, 'duration_secs': 0.012875} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.955417] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24587e88-dde5-4704-a6c9-fd35eb14d48e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.962442] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 951.962442] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52175581-ab1f-2809-690e-6c1fdac37cc2" [ 951.962442] env[61911]: _type = "Task" [ 951.962442] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.965284] env[61911]: DEBUG nova.network.neutron [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [{"id": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "address": "fa:16:3e:c5:e0:b6", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce08c90-28", "ovs_interfaceid": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.972279] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52175581-ab1f-2809-690e-6c1fdac37cc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.055648] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251394, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.127586] env[61911]: DEBUG nova.network.neutron [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 952.135971] env[61911]: INFO nova.compute.manager [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Took 20.69 seconds to build instance. [ 952.176024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.176024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.176024] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 952.376446] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c30c6e37-995d-46fa-8558-852cd2d5c959 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "8d9881fe-4bb5-4d85-9be9-32c2eca32a9a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.063s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.383809] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.384098] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.385679] env[61911]: INFO nova.compute.claims [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.436047] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251393, 'name': ReconfigVM_Task, 'duration_secs': 0.755737} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.436413] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d/8051e8ec-6d82-416c-858a-f0cdd00f869d.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.437109] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-618aa271-95ab-47c8-b79d-bdf0bf7673b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.441190] env[61911]: DEBUG nova.network.neutron [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Updating instance_info_cache with network_info: [{"id": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "address": "fa:16:3e:a4:2b:ca", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33e8e080-83", "ovs_interfaceid": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.450379] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 952.450379] env[61911]: value = "task-1251395" [ 952.450379] env[61911]: _type = "Task" [ 952.450379] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.463914] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251395, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.467779] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Releasing lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.468580] env[61911]: DEBUG nova.objects.instance [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lazy-loading 'flavor' on Instance uuid 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.476141] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52175581-ab1f-2809-690e-6c1fdac37cc2, 'name': SearchDatastore_Task, 'duration_secs': 0.043883} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.476344] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.476613] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] ec53a8ee-df42-4003-a933-17aa4c90c3aa/ec53a8ee-df42-4003-a933-17aa4c90c3aa.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 952.476948] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-98bdb196-905e-44ee-8518-51867f96fbf7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.486211] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 952.486211] env[61911]: value = "task-1251396" [ 952.486211] env[61911]: _type = "Task" [ 952.486211] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.496076] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251396, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.554830] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251394, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.647301] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99977247-dca4-4c6a-886a-08e41f879b46 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.206s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.727150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "75577b68-b012-43d5-abdb-bd5f54c249d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.727150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.727150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "75577b68-b012-43d5-abdb-bd5f54c249d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.727150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.727150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.728025] env[61911]: INFO nova.compute.manager [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Terminating instance [ 952.859890] env[61911]: DEBUG nova.compute.manager [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Received event network-vif-plugged-33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 952.860133] env[61911]: DEBUG oslo_concurrency.lockutils [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] Acquiring lock "fddba36d-1b15-43fb-9e99-68880e8a235c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.860354] env[61911]: DEBUG oslo_concurrency.lockutils [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.860567] env[61911]: DEBUG oslo_concurrency.lockutils [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.860784] env[61911]: DEBUG nova.compute.manager [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] No waiting events found dispatching network-vif-plugged-33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 952.861368] env[61911]: WARNING nova.compute.manager [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Received unexpected event network-vif-plugged-33e8e080-830e-4a6a-8e9b-39eded7bf6de for instance with vm_state building and task_state spawning. [ 952.861368] env[61911]: DEBUG nova.compute.manager [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Received event network-changed-33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 952.861644] env[61911]: DEBUG nova.compute.manager [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Refreshing instance network info cache due to event network-changed-33e8e080-830e-4a6a-8e9b-39eded7bf6de. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 952.861644] env[61911]: DEBUG oslo_concurrency.lockutils [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] Acquiring lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.944100] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.944453] env[61911]: DEBUG nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Instance network_info: |[{"id": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "address": "fa:16:3e:a4:2b:ca", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33e8e080-83", "ovs_interfaceid": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 952.944810] env[61911]: DEBUG oslo_concurrency.lockutils [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] Acquired lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.945077] env[61911]: DEBUG nova.network.neutron [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Refreshing network info cache for port 33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.946362] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:2b:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33e8e080-830e-4a6a-8e9b-39eded7bf6de', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.954890] env[61911]: DEBUG oslo.service.loopingcall [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.956133] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.959595] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b352cd43-c11c-4153-a67a-778424fdd034 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.980147] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fd01ad-b4ab-4ebd-a767-3e291e26b689 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.992924] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.992924] env[61911]: value = "task-1251397" [ 952.992924] env[61911]: _type = "Task" [ 952.992924] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.993284] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251395, 'name': Rename_Task, 'duration_secs': 0.254276} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.998826] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.030176] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e805e0c7-72e7-42ed-858b-95bb5665e4ed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.033724] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.034695] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6c26d00-298d-4cea-a698-d3e3027f4314 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.041965] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251396, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.045804] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251397, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.051781] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 953.051781] env[61911]: value = "task-1251398" [ 953.051781] env[61911]: _type = "Task" [ 953.051781] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.051781] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 953.051781] env[61911]: value = "task-1251399" [ 953.051781] env[61911]: _type = "Task" [ 953.051781] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.062907] env[61911]: DEBUG oslo_vmware.api [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251394, 'name': RemoveSnapshot_Task, 'duration_secs': 1.492304} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.063520] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 953.073384] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251398, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.073845] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251399, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.234418] env[61911]: DEBUG nova.compute.manager [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 953.234855] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.235913] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762169fb-9682-420a-aa27-333a80b26d69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.244676] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.244949] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-462f59cb-0112-4da1-a9a3-9c13f2e6ca98 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.254403] env[61911]: DEBUG oslo_vmware.api [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 953.254403] env[61911]: value = "task-1251400" [ 953.254403] env[61911]: _type = "Task" [ 953.254403] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.270208] env[61911]: DEBUG oslo_vmware.api [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.507613] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251396, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738562} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.511779] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] ec53a8ee-df42-4003-a933-17aa4c90c3aa/ec53a8ee-df42-4003-a933-17aa4c90c3aa.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.512095] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.512430] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251397, 'name': CreateVM_Task, 'duration_secs': 0.494748} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.512593] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-95b00334-172f-4dba-9315-2d4470cfa9de {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.514408] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.519682] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.519901] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.521943] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.524026] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bfb06c9-df7c-4eca-b972-b39e42170395 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.528588] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 953.528588] env[61911]: value = "task-1251401" [ 953.528588] env[61911]: _type = "Task" [ 953.528588] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.528888] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 953.528888] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221e58f-7654-2291-fa26-e85b802b5549" [ 953.528888] env[61911]: _type = "Task" [ 953.528888] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.540984] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221e58f-7654-2291-fa26-e85b802b5549, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.541940] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.547587] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251401, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.573754] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251399, 'name': PowerOffVM_Task, 'duration_secs': 0.410685} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.574040] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251398, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.574358] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.579899] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfiguring VM instance instance-0000004d to detach disk 2002 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 953.583876] env[61911]: WARNING nova.compute.manager [None req-a342b570-cd08-4cea-997f-1a3c69fa56a3 tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Image not found during snapshot: nova.exception.ImageNotFound: Image 0127f673-151a-45e0-aedf-e4b30b4dcf9d could not be found. [ 953.585138] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3748ede0-0b57-440f-9561-14cc6c5c62c2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.606845] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 953.606845] env[61911]: value = "task-1251402" [ 953.606845] env[61911]: _type = "Task" [ 953.606845] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.616342] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251402, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.679591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba847791-789a-4fbf-9a70-92aa85e1ab68 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.688772] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b96bfc-039f-496e-83ce-4767c664d845 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.718328] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0345d22-ac1f-4aa3-9692-db9deef01602 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.725219] env[61911]: DEBUG nova.network.neutron [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Updated VIF entry in instance network info cache for port 33e8e080-830e-4a6a-8e9b-39eded7bf6de. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.725565] env[61911]: DEBUG nova.network.neutron [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Updating instance_info_cache with network_info: [{"id": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "address": "fa:16:3e:a4:2b:ca", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33e8e080-83", "ovs_interfaceid": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.730017] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6980892-1860-4cbf-ade5-6a903d857fe1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.741080] env[61911]: DEBUG nova.compute.provider_tree [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.767127] env[61911]: DEBUG oslo_vmware.api [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251400, 'name': PowerOffVM_Task, 'duration_secs': 0.204467} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.767498] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.767744] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.768018] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc1de92d-6e00-42ed-8e8f-60a1d15b28b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.833183] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.833425] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.833774] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleting the datastore file [datastore2] 75577b68-b012-43d5-abdb-bd5f54c249d5 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.833870] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86bfda56-2fb1-4634-9cf9-256f0086d4a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.845024] env[61911]: DEBUG oslo_vmware.api [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 953.845024] env[61911]: value = "task-1251404" [ 953.845024] env[61911]: _type = "Task" [ 953.845024] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.852328] env[61911]: DEBUG oslo_vmware.api [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.041570] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251401, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.153982} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.045489] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.045826] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221e58f-7654-2291-fa26-e85b802b5549, 'name': SearchDatastore_Task, 'duration_secs': 0.012312} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.046552] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d25faf-8e4e-4d83-b1ea-930ffc40b771 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.049226] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.049458] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.049690] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.049690] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.050020] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.050480] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.050665] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 954.050855] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7851d3b3-02a9-45af-8f1c-3b55f77453b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.052704] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.053458] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.065447] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.075889] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] ec53a8ee-df42-4003-a933-17aa4c90c3aa/ec53a8ee-df42-4003-a933-17aa4c90c3aa.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.078402] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.079135] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8316a9b1-9beb-445e-85b6-6d54edeeb38a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.096317] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.097024] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.103881] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.104845] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eb74984-d1f2-421f-9ea9-04c3b938c397 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.106912] env[61911]: DEBUG oslo_vmware.api [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251398, 'name': PowerOnVM_Task, 'duration_secs': 0.773901} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.107388] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.107540] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 954.108559] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.110750] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.121125] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 954.121125] env[61911]: value = "task-1251405" [ 954.121125] env[61911]: _type = "Task" [ 954.121125] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.121396] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 954.121396] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e9ec7a-ca9e-1814-0fac-101ed6a8ecbf" [ 954.121396] env[61911]: _type = "Task" [ 954.121396] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.135022] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251402, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.143369] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251405, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.145343] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e9ec7a-ca9e-1814-0fac-101ed6a8ecbf, 'name': SearchDatastore_Task, 'duration_secs': 0.011392} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.146182] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ffadb80-02ef-4501-9813-f250df1bb248 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.152521] env[61911]: DEBUG nova.compute.manager [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Stashing vm_state: active {{(pid=61911) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 954.156955] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 954.156955] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ec4acf-0884-2d66-b7c9-642a61e14594" [ 954.156955] env[61911]: _type = "Task" [ 954.156955] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.167141] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ec4acf-0884-2d66-b7c9-642a61e14594, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.233797] env[61911]: DEBUG oslo_concurrency.lockutils [req-90b53274-d479-43de-a6af-7837719c86a8 req-9e4cf6e2-7106-4326-b1ce-1306284b64fb service nova] Releasing lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.241296] env[61911]: DEBUG nova.compute.manager [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 954.242139] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab3e3b4-fc48-421f-b43c-5cd108f445f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.245529] env[61911]: DEBUG nova.scheduler.client.report [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.353371] env[61911]: DEBUG oslo_vmware.api [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329964} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.353657] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.353842] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.354038] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.354221] env[61911]: INFO nova.compute.manager [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 954.354462] env[61911]: DEBUG oslo.service.loopingcall [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.354663] env[61911]: DEBUG nova.compute.manager [-] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 954.354759] env[61911]: DEBUG nova.network.neutron [-] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.608145] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.608401] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.608958] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.608958] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.608958] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.614582] env[61911]: INFO nova.compute.manager [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Terminating instance [ 954.616745] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.624128] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251402, 'name': ReconfigVM_Task, 'duration_secs': 0.55824} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.627562] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfigured VM instance instance-0000004d to detach disk 2002 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 954.627780] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.628056] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7384d0ed-0495-423c-badb-e8d1d5783614 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.636271] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251405, 'name': ReconfigVM_Task, 'duration_secs': 0.37815} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.637575] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Reconfigured VM instance instance-00000056 to attach disk [datastore1] ec53a8ee-df42-4003-a933-17aa4c90c3aa/ec53a8ee-df42-4003-a933-17aa4c90c3aa.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.638394] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 954.638394] env[61911]: value = "task-1251406" [ 954.638394] env[61911]: _type = "Task" [ 954.638394] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.638724] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3dcb5f0a-bc17-49e4-9f97-3ad33c3b8224 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.649664] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.650333] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 954.650333] env[61911]: value = "task-1251407" [ 954.650333] env[61911]: _type = "Task" [ 954.650333] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.662888] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251407, 'name': Rename_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.674659] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ec4acf-0884-2d66-b7c9-642a61e14594, 'name': SearchDatastore_Task, 'duration_secs': 0.013529} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.676040] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.676040] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] fddba36d-1b15-43fb-9e99-68880e8a235c/fddba36d-1b15-43fb-9e99-68880e8a235c.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.676040] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5c99c21-0acd-4870-8881-37bd106f6541 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.679218] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.682407] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 954.682407] env[61911]: value = "task-1251408" [ 954.682407] env[61911]: _type = "Task" [ 954.682407] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.693149] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.751608] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.752513] env[61911]: DEBUG nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 954.755185] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.138s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.755374] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.755532] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 954.755860] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.077s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.759017] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4634fda-79db-4269-8906-d23755eeefd0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.768219] env[61911]: DEBUG oslo_concurrency.lockutils [None req-591f7d54-3c7d-44d1-949d-0f970d5ee3f8 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 30.567s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.776231] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9f25dd-bc76-43e2-b399-6181854d5d8e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.792224] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c565410-6887-461b-a5fc-07fcf68298f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.800475] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9545762-2cb3-493c-9b5d-53c60be2560d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.832619] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179215MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 954.832795] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.884321] env[61911]: DEBUG nova.compute.manager [req-d4caf758-3225-46f1-9dab-8212c9e58d2f req-a4d28e49-6bc6-409a-ae24-dcc7313e544f service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Received event network-vif-deleted-4da57eb9-f431-463c-b1fc-292ec04a8cac {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 954.884524] env[61911]: INFO nova.compute.manager [req-d4caf758-3225-46f1-9dab-8212c9e58d2f req-a4d28e49-6bc6-409a-ae24-dcc7313e544f service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Neutron deleted interface 4da57eb9-f431-463c-b1fc-292ec04a8cac; detaching it from the instance and deleting it from the info cache [ 954.884715] env[61911]: DEBUG nova.network.neutron [req-d4caf758-3225-46f1-9dab-8212c9e58d2f req-a4d28e49-6bc6-409a-ae24-dcc7313e544f service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.120014] env[61911]: DEBUG nova.compute.manager [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 955.123031] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.123031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202e9ee9-b454-46d7-87fd-8a00339711ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.137166] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.137166] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c64c978c-187e-4da8-9b59-4486676d0071 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.150508] env[61911]: DEBUG oslo_vmware.api [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 955.150508] env[61911]: value = "task-1251409" [ 955.150508] env[61911]: _type = "Task" [ 955.150508] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.170739] env[61911]: DEBUG nova.network.neutron [-] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.171424] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251406, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.184474] env[61911]: DEBUG oslo_vmware.api [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251409, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.184474] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251407, 'name': Rename_Task, 'duration_secs': 0.202017} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.194451] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 955.194451] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58db08d7-2946-472f-89d3-6da9464bcf5f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.199496] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251408, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.201654] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 955.201654] env[61911]: value = "task-1251410" [ 955.201654] env[61911]: _type = "Task" [ 955.201654] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.212178] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.266105] env[61911]: INFO nova.compute.claims [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.269868] env[61911]: DEBUG nova.compute.utils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.272534] env[61911]: DEBUG nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 955.274024] env[61911]: DEBUG nova.network.neutron [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 955.337214] env[61911]: DEBUG nova.policy [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c6995470637a4f159572c852822a7266', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78e06816d3f54611bc3224aa5085d3c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.391024] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c46a26ed-4c3d-488d-9d14-9a12f544f2db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.398234] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d36cd5-fb5b-4ca7-a63b-f97385fc8e2f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.435296] env[61911]: DEBUG nova.compute.manager [req-d4caf758-3225-46f1-9dab-8212c9e58d2f req-a4d28e49-6bc6-409a-ae24-dcc7313e544f service nova] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Detach interface failed, port_id=4da57eb9-f431-463c-b1fc-292ec04a8cac, reason: Instance 75577b68-b012-43d5-abdb-bd5f54c249d5 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 955.633863] env[61911]: DEBUG nova.network.neutron [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Successfully created port: a3f9143d-1bd7-42d4-be27-f2a515d78820 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.654020] env[61911]: DEBUG oslo_vmware.api [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251406, 'name': PowerOnVM_Task, 'duration_secs': 0.677795} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.654020] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.654020] env[61911]: DEBUG nova.compute.manager [None req-d8ce6d41-3468-4fbe-ada7-cb62cde5b4b3 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 955.655314] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c911de-d95c-4de6-9a84-6646e03b7044 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.663238] env[61911]: DEBUG oslo_vmware.api [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251409, 'name': PowerOffVM_Task, 'duration_secs': 0.206596} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.665294] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.667031] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.670147] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74269e21-9c9b-425a-9e03-f0cd9224e3ad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.676051] env[61911]: INFO nova.compute.manager [-] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Took 1.32 seconds to deallocate network for instance. [ 955.693710] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.711844} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.694380] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] fddba36d-1b15-43fb-9e99-68880e8a235c/fddba36d-1b15-43fb-9e99-68880e8a235c.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.694906] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.695463] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ebc0475-d98b-4ed3-8f05-df318b1568d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.708711] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 955.708711] env[61911]: value = "task-1251412" [ 955.708711] env[61911]: _type = "Task" [ 955.708711] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.719124] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251410, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.724223] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251412, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.748098] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.748406] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.748635] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleting the datastore file [datastore1] 3c4c8943-5324-4c3e-b7e0-7ea5972bc026 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.748994] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80d874a3-155b-4b00-8414-6dff8abf4515 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.755478] env[61911]: DEBUG oslo_vmware.api [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for the task: (returnval){ [ 955.755478] env[61911]: value = "task-1251413" [ 955.755478] env[61911]: _type = "Task" [ 955.755478] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.763598] env[61911]: DEBUG oslo_vmware.api [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.777375] env[61911]: DEBUG nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 955.782054] env[61911]: INFO nova.compute.resource_tracker [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating resource usage from migration 2e712027-ffdf-4233-8dba-1d056402fa73 [ 955.976693] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f029ff5e-19f4-4680-a6fc-46359bde4b29 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.985048] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2099b3-71d6-40c5-bd37-01d4a2db69b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.016165] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6f5fe1-fcdb-4169-b4a1-e12444db2c09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.023282] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6194269-1c8a-439b-8a66-f06b7cfe9c2a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.036304] env[61911]: DEBUG nova.compute.provider_tree [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.181894] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.219282] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251410, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.226389] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251412, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112344} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.226858] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.227854] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6252ddd9-f799-4a8e-ac0f-757f93a75b6f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.252602] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] fddba36d-1b15-43fb-9e99-68880e8a235c/fddba36d-1b15-43fb-9e99-68880e8a235c.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.253000] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb49df4c-b6fa-4390-bbd8-23c9abcd2c1c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.279850] env[61911]: DEBUG oslo_vmware.api [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.281486] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 956.281486] env[61911]: value = "task-1251414" [ 956.281486] env[61911]: _type = "Task" [ 956.281486] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.293400] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251414, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.539668] env[61911]: DEBUG nova.scheduler.client.report [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 956.713905] env[61911]: DEBUG oslo_vmware.api [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251410, 'name': PowerOnVM_Task, 'duration_secs': 1.438178} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.714474] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.714738] env[61911]: INFO nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Took 8.95 seconds to spawn the instance on the hypervisor. [ 956.714963] env[61911]: DEBUG nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 956.715783] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6248f71c-d81d-471a-a1c6-d7738ffbacf7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.782728] env[61911]: DEBUG oslo_vmware.api [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Task: {'id': task-1251413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.577327} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.783475] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.783860] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.786423] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.786674] env[61911]: INFO nova.compute.manager [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Took 1.67 seconds to destroy the instance on the hypervisor. [ 956.786980] env[61911]: DEBUG oslo.service.loopingcall [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.790173] env[61911]: DEBUG nova.compute.manager [-] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 956.790292] env[61911]: DEBUG nova.network.neutron [-] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.792889] env[61911]: DEBUG nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 956.801129] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.817982] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 956.818277] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.818446] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 956.818627] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.818776] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 956.818924] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 956.819304] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 956.819528] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 956.819758] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 956.819969] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 956.820204] env[61911]: DEBUG nova.virt.hardware [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 956.821212] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5073f7-ea1d-4c73-bb7c-acdd8392cf2d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.830377] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70c367f-25f8-4206-8238-47bfdb6f39e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.893947] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.895032] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.997800] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9283e0-e2c8-4cc0-bdbb-7490994c5a28 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.007401] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a8406-ba1a-4527-a69e-8ec87e6ee0c1 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Suspending the VM {{(pid=61911) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 957.007401] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6f59ee52-0966-4e5d-9ece-ba38e781ecca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.014288] env[61911]: DEBUG oslo_vmware.api [None req-0f9a8406-ba1a-4527-a69e-8ec87e6ee0c1 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 957.014288] env[61911]: value = "task-1251415" [ 957.014288] env[61911]: _type = "Task" [ 957.014288] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.023302] env[61911]: DEBUG oslo_vmware.api [None req-0f9a8406-ba1a-4527-a69e-8ec87e6ee0c1 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251415, 'name': SuspendVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.048204] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.290s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.048204] env[61911]: INFO nova.compute.manager [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Migrating [ 957.055581] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.223s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.173650] env[61911]: DEBUG nova.network.neutron [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Successfully updated port: a3f9143d-1bd7-42d4-be27-f2a515d78820 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.238787] env[61911]: INFO nova.compute.manager [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Took 23.61 seconds to build instance. [ 957.294420] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251414, 'name': ReconfigVM_Task, 'duration_secs': 0.873628} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.296161] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Reconfigured VM instance instance-00000057 to attach disk [datastore2] fddba36d-1b15-43fb-9e99-68880e8a235c/fddba36d-1b15-43fb-9e99-68880e8a235c.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.296161] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cde433f3-a1c1-45a1-95a3-294674d98069 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.302868] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 957.302868] env[61911]: value = "task-1251416" [ 957.302868] env[61911]: _type = "Task" [ 957.302868] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.314248] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251416, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.350323] env[61911]: DEBUG nova.compute.manager [req-e14816f8-0b38-4f4c-9fad-e36df74e411b req-b0d3cc28-141e-411a-a3a5-09e653272b89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Received event network-vif-plugged-a3f9143d-1bd7-42d4-be27-f2a515d78820 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 957.350640] env[61911]: DEBUG oslo_concurrency.lockutils [req-e14816f8-0b38-4f4c-9fad-e36df74e411b req-b0d3cc28-141e-411a-a3a5-09e653272b89 service nova] Acquiring lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.350837] env[61911]: DEBUG oslo_concurrency.lockutils [req-e14816f8-0b38-4f4c-9fad-e36df74e411b req-b0d3cc28-141e-411a-a3a5-09e653272b89 service nova] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.351034] env[61911]: DEBUG oslo_concurrency.lockutils [req-e14816f8-0b38-4f4c-9fad-e36df74e411b req-b0d3cc28-141e-411a-a3a5-09e653272b89 service nova] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.351221] env[61911]: DEBUG nova.compute.manager [req-e14816f8-0b38-4f4c-9fad-e36df74e411b req-b0d3cc28-141e-411a-a3a5-09e653272b89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] No waiting events found dispatching network-vif-plugged-a3f9143d-1bd7-42d4-be27-f2a515d78820 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 957.351422] env[61911]: WARNING nova.compute.manager [req-e14816f8-0b38-4f4c-9fad-e36df74e411b req-b0d3cc28-141e-411a-a3a5-09e653272b89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Received unexpected event network-vif-plugged-a3f9143d-1bd7-42d4-be27-f2a515d78820 for instance with vm_state building and task_state spawning. [ 957.397092] env[61911]: DEBUG nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 957.425969] env[61911]: DEBUG nova.compute.manager [req-18fc6517-2d17-476f-8a27-8cd906b2686a req-596d133c-95a5-4e21-85a0-ebbb988e0471 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Received event network-vif-deleted-38e33f6e-614a-4041-a8d0-136e96470621 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 957.426276] env[61911]: INFO nova.compute.manager [req-18fc6517-2d17-476f-8a27-8cd906b2686a req-596d133c-95a5-4e21-85a0-ebbb988e0471 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Neutron deleted interface 38e33f6e-614a-4041-a8d0-136e96470621; detaching it from the instance and deleting it from the info cache [ 957.426519] env[61911]: DEBUG nova.network.neutron [req-18fc6517-2d17-476f-8a27-8cd906b2686a req-596d133c-95a5-4e21-85a0-ebbb988e0471 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.525240] env[61911]: DEBUG oslo_vmware.api [None req-0f9a8406-ba1a-4527-a69e-8ec87e6ee0c1 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251415, 'name': SuspendVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.578021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.578021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.578021] env[61911]: DEBUG nova.network.neutron [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.638236] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.638236] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.677709] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "refresh_cache-7060c2a1-7546-41f9-891d-af6a59fa1aa5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.678474] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquired lock "refresh_cache-7060c2a1-7546-41f9-891d-af6a59fa1aa5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.684907] env[61911]: DEBUG nova.network.neutron [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.741142] env[61911]: DEBUG oslo_concurrency.lockutils [None req-71cd7f18-79ed-4a15-a14f-835de110b8f2 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.123s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.763695] env[61911]: DEBUG nova.network.neutron [-] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.814639] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251416, 'name': Rename_Task, 'duration_secs': 0.208314} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.814927] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.815308] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c89d5467-44ca-418f-851e-df1b6bf35347 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.827237] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 957.827237] env[61911]: value = "task-1251417" [ 957.827237] env[61911]: _type = "Task" [ 957.827237] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.835490] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251417, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.919453] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.929434] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0686cc42-e3ef-4524-b7ab-9e5734a3cdec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.944195] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927b1f05-99e9-48eb-b419-2bfab3f26f00 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.973313] env[61911]: DEBUG nova.compute.manager [req-18fc6517-2d17-476f-8a27-8cd906b2686a req-596d133c-95a5-4e21-85a0-ebbb988e0471 service nova] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Detach interface failed, port_id=38e33f6e-614a-4041-a8d0-136e96470621, reason: Instance 3c4c8943-5324-4c3e-b7e0-7ea5972bc026 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 958.025116] env[61911]: DEBUG oslo_vmware.api [None req-0f9a8406-ba1a-4527-a69e-8ec87e6ee0c1 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251415, 'name': SuspendVM_Task, 'duration_secs': 0.732429} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.025451] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9a8406-ba1a-4527-a69e-8ec87e6ee0c1 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Suspended the VM {{(pid=61911) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 958.025643] env[61911]: DEBUG nova.compute.manager [None req-0f9a8406-ba1a-4527-a69e-8ec87e6ee0c1 tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 958.026480] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1143a123-a507-4557-884f-81cb60d3c1e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.081789] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Applying migration context for instance f386424d-953d-431d-a56b-8542f2a3458c as it has an incoming, in-progress migration 2e712027-ffdf-4233-8dba-1d056402fa73. Migration status is pre-migrating {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 958.089210] env[61911]: INFO nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating resource usage from migration 2e712027-ffdf-4233-8dba-1d056402fa73 [ 958.090414] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.095089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.095089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.095089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.095089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.095353] env[61911]: INFO nova.compute.manager [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Terminating instance [ 958.110301] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 31f0a4f7-a69d-4ec4-966d-93f402163b88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.110466] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bafd58bc-8ae7-49a4-a039-31d328699010 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.110684] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance c95707ca-7e32-42c0-914a-f0b178f07997 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.110740] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5b7d09e1-ee14-4001-8d73-14e763402670 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.110844] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.110989] env[61911]: WARNING nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 75577b68-b012-43d5-abdb-bd5f54c249d5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 958.111156] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8051e8ec-6d82-416c-858a-f0cdd00f869d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.111234] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 3c4c8943-5324-4c3e-b7e0-7ea5972bc026 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.111454] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance ec53a8ee-df42-4003-a933-17aa4c90c3aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.111454] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance fddba36d-1b15-43fb-9e99-68880e8a235c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.111558] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 7060c2a1-7546-41f9-891d-af6a59fa1aa5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.111658] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Migration 2e712027-ffdf-4233-8dba-1d056402fa73 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 958.111767] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance f386424d-953d-431d-a56b-8542f2a3458c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 958.140412] env[61911]: DEBUG nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 958.224278] env[61911]: DEBUG nova.network.neutron [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.267876] env[61911]: INFO nova.compute.manager [-] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Took 1.48 seconds to deallocate network for instance. [ 958.303812] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.303812] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.335462] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251417, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.372722] env[61911]: DEBUG nova.network.neutron [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance_info_cache with network_info: [{"id": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "address": "fa:16:3e:ef:fb:2e", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5937b7a1-04", "ovs_interfaceid": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.377656] env[61911]: DEBUG nova.network.neutron [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Updating instance_info_cache with network_info: [{"id": "a3f9143d-1bd7-42d4-be27-f2a515d78820", "address": "fa:16:3e:65:5f:f4", "network": {"id": "b5baa01c-13d0-4eac-b210-7c3a64847094", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1829313513-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78e06816d3f54611bc3224aa5085d3c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3f9143d-1b", "ovs_interfaceid": "a3f9143d-1bd7-42d4-be27-f2a515d78820", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.601990] env[61911]: DEBUG nova.compute.manager [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 958.602234] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.603642] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c7fe55-86af-42cf-9236-7cc42083f500 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.612083] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.612313] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07c3909f-97a6-4e98-8a72-57ac9582aa8b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.615462] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b1d2b205-81d7-410b-901d-ccd608bc5bec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 958.619040] env[61911]: DEBUG oslo_vmware.api [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 958.619040] env[61911]: value = "task-1251418" [ 958.619040] env[61911]: _type = "Task" [ 958.619040] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.628646] env[61911]: DEBUG oslo_vmware.api [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251418, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.662794] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.781758] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.805562] env[61911]: DEBUG nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 958.836869] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251417, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.877428] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.879782] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Releasing lock "refresh_cache-7060c2a1-7546-41f9-891d-af6a59fa1aa5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.879782] env[61911]: DEBUG nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Instance network_info: |[{"id": "a3f9143d-1bd7-42d4-be27-f2a515d78820", "address": "fa:16:3e:65:5f:f4", "network": {"id": "b5baa01c-13d0-4eac-b210-7c3a64847094", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1829313513-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78e06816d3f54611bc3224aa5085d3c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3f9143d-1b", "ovs_interfaceid": "a3f9143d-1bd7-42d4-be27-f2a515d78820", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 958.880526] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:5f:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3f9143d-1bd7-42d4-be27-f2a515d78820', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.888505] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Creating folder: Project (78e06816d3f54611bc3224aa5085d3c2). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.889011] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6347cc7-7bbd-4590-8c49-d8d71b8f5979 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.900608] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Created folder: Project (78e06816d3f54611bc3224aa5085d3c2) in parent group-v269521. [ 958.900842] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Creating folder: Instances. Parent ref: group-v269643. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.901119] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9d3fbee-ed5c-4d46-91be-1cb446893f18 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.910033] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Created folder: Instances in parent group-v269643. [ 958.910276] env[61911]: DEBUG oslo.service.loopingcall [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.910467] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.910674] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dc1769a-2490-416e-b1a3-dd1016bd2157 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.930086] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.930086] env[61911]: value = "task-1251421" [ 958.930086] env[61911]: _type = "Task" [ 958.930086] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.938290] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251421, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.119042] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 959.133031] env[61911]: DEBUG oslo_vmware.api [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251418, 'name': PowerOffVM_Task, 'duration_secs': 0.283481} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.134225] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.134467] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.134798] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3752d6e-fbb4-49f6-9a64-f29a4c8bcaf1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.196348] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.196617] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.196808] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleting the datastore file [datastore1] ec53a8ee-df42-4003-a933-17aa4c90c3aa {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.197071] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12b30271-03ad-47de-92bd-ccc3e875574a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.203722] env[61911]: DEBUG oslo_vmware.api [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 959.203722] env[61911]: value = "task-1251423" [ 959.203722] env[61911]: _type = "Task" [ 959.203722] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.211232] env[61911]: DEBUG oslo_vmware.api [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.326187] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.336748] env[61911]: DEBUG oslo_vmware.api [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251417, 'name': PowerOnVM_Task, 'duration_secs': 1.211923} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.337753] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.337753] env[61911]: INFO nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Took 9.18 seconds to spawn the instance on the hypervisor. [ 959.337753] env[61911]: DEBUG nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 959.338249] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b40bc37-fb95-47a8-8e2f-daff2d17dd85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.392649] env[61911]: DEBUG nova.compute.manager [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Received event network-changed-a3f9143d-1bd7-42d4-be27-f2a515d78820 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 959.392858] env[61911]: DEBUG nova.compute.manager [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Refreshing instance network info cache due to event network-changed-a3f9143d-1bd7-42d4-be27-f2a515d78820. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 959.393097] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Acquiring lock "refresh_cache-7060c2a1-7546-41f9-891d-af6a59fa1aa5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.393249] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Acquired lock "refresh_cache-7060c2a1-7546-41f9-891d-af6a59fa1aa5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.393412] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Refreshing network info cache for port a3f9143d-1bd7-42d4-be27-f2a515d78820 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.440537] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251421, 'name': CreateVM_Task, 'duration_secs': 0.398606} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.440726] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.441477] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.441635] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.442203] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.442282] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2870dd2-5a57-4284-9c19-3cc6d98feaa0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.447172] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 959.447172] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ab7203-9c49-870a-4e73-58888318171e" [ 959.447172] env[61911]: _type = "Task" [ 959.447172] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.455098] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ab7203-9c49-870a-4e73-58888318171e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.627109] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance d9234a5f-4d26-4bdc-8f32-3120830a8abe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 959.627109] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 959.627340] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 959.712488] env[61911]: DEBUG oslo_vmware.api [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171477} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.714805] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.715017] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.715223] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.715418] env[61911]: INFO nova.compute.manager [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Took 1.11 seconds to destroy the instance on the hypervisor. [ 959.715655] env[61911]: DEBUG oslo.service.loopingcall [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.716027] env[61911]: DEBUG nova.compute.manager [-] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 959.716158] env[61911]: DEBUG nova.network.neutron [-] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 959.823225] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d33591-f3c5-4d9a-9983-b057e70d757a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.830945] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80046d9-e73a-4dc3-96b8-48b34beb80ca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.865076] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3adbac5-1914-4c0e-98f4-7d62c837cacf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.870738] env[61911]: INFO nova.compute.manager [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Took 24.22 seconds to build instance. [ 959.878315] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed36528-811c-439e-98e3-caf45ea1d949 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.902549] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.957770] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ab7203-9c49-870a-4e73-58888318171e, 'name': SearchDatastore_Task, 'duration_secs': 0.010155} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.960024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.960270] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.960503] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.960651] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.960853] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.961281] env[61911]: INFO nova.compute.manager [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Resuming [ 959.961790] env[61911]: DEBUG nova.objects.instance [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lazy-loading 'flavor' on Instance uuid 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.963013] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6335c86a-95d1-4a0c-9ed1-2348948d68af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.973273] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.973491] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.974251] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91f5fb48-2df0-47d8-8b70-a57ff09f74f4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.980159] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 959.980159] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b3e1c-63bb-591c-226b-6151d22cfc62" [ 959.980159] env[61911]: _type = "Task" [ 959.980159] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.988673] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b3e1c-63bb-591c-226b-6151d22cfc62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.243678] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Updated VIF entry in instance network info cache for port a3f9143d-1bd7-42d4-be27-f2a515d78820. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 960.244122] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Updating instance_info_cache with network_info: [{"id": "a3f9143d-1bd7-42d4-be27-f2a515d78820", "address": "fa:16:3e:65:5f:f4", "network": {"id": "b5baa01c-13d0-4eac-b210-7c3a64847094", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1829313513-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78e06816d3f54611bc3224aa5085d3c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3f9143d-1b", "ovs_interfaceid": "a3f9143d-1bd7-42d4-be27-f2a515d78820", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.373420] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b2b3af2e-da4b-4bb5-a019-178ac636ef14 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.727s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.404885] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613982d6-c285-4da2-ae67-ed065ad40c85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.408592] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 960.427684] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance 'f386424d-953d-431d-a56b-8542f2a3458c' progress to 0 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 960.496325] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b3e1c-63bb-591c-226b-6151d22cfc62, 'name': SearchDatastore_Task, 'duration_secs': 0.009571} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.497650] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2289004f-77a6-4f08-ae2a-a2f245705b1d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.504369] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 960.504369] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52148f82-8731-2fa5-904a-bfff654521ec" [ 960.504369] env[61911]: _type = "Task" [ 960.504369] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.513017] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52148f82-8731-2fa5-904a-bfff654521ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.657836] env[61911]: DEBUG nova.network.neutron [-] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.746973] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Releasing lock "refresh_cache-7060c2a1-7546-41f9-891d-af6a59fa1aa5" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.747226] env[61911]: DEBUG nova.compute.manager [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Received event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 960.747570] env[61911]: DEBUG nova.compute.manager [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing instance network info cache due to event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 960.747975] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Acquiring lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.748264] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Acquired lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.748589] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing network info cache for port 7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.913780] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 960.914109] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.859s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.914332] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.733s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.914523] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.916722] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.997s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.918177] env[61911]: INFO nova.compute.claims [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.933213] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.935734] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59700f1d-2061-473f-989b-3c3bd281dc97 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.942937] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 960.942937] env[61911]: value = "task-1251424" [ 960.942937] env[61911]: _type = "Task" [ 960.942937] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.949226] env[61911]: INFO nova.scheduler.client.report [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted allocations for instance 75577b68-b012-43d5-abdb-bd5f54c249d5 [ 960.959892] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251424, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.016702] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52148f82-8731-2fa5-904a-bfff654521ec, 'name': SearchDatastore_Task, 'duration_secs': 0.020861} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.016887] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.017169] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 7060c2a1-7546-41f9-891d-af6a59fa1aa5/7060c2a1-7546-41f9-891d-af6a59fa1aa5.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 961.017426] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d94215d8-dd2d-4ff8-ba53-a3a4e212fcac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.024233] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 961.024233] env[61911]: value = "task-1251425" [ 961.024233] env[61911]: _type = "Task" [ 961.024233] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.032083] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.162694] env[61911]: INFO nova.compute.manager [-] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Took 1.45 seconds to deallocate network for instance. [ 961.434912] env[61911]: DEBUG nova.compute.manager [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Received event network-vif-deleted-def5852e-6e08-47c9-9179-88d9b4161e8d {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 961.435158] env[61911]: DEBUG nova.compute.manager [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Received event network-changed-33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 961.435859] env[61911]: DEBUG nova.compute.manager [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Refreshing instance network info cache due to event network-changed-33e8e080-830e-4a6a-8e9b-39eded7bf6de. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 961.436340] env[61911]: DEBUG oslo_concurrency.lockutils [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] Acquiring lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.437265] env[61911]: DEBUG oslo_concurrency.lockutils [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] Acquired lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.437605] env[61911]: DEBUG nova.network.neutron [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Refreshing network info cache for port 33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 961.454835] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251424, 'name': PowerOffVM_Task, 'duration_secs': 0.16518} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.454835] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.454990] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance 'f386424d-953d-431d-a56b-8542f2a3458c' progress to 17 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 961.466184] env[61911]: DEBUG oslo_concurrency.lockutils [None req-066c678f-7005-4a9f-8ef4-4733501fb9e1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "75577b68-b012-43d5-abdb-bd5f54c249d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.741s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.475482] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.475686] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquired lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.475882] env[61911]: DEBUG nova.network.neutron [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.535660] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251425, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468883} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.535660] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 7060c2a1-7546-41f9-891d-af6a59fa1aa5/7060c2a1-7546-41f9-891d-af6a59fa1aa5.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.536098] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.536524] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1b9d12e-3e06-40ee-94b2-4f1930be3c43 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.546120] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 961.546120] env[61911]: value = "task-1251426" [ 961.546120] env[61911]: _type = "Task" [ 961.546120] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.552231] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251426, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.662476] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updated VIF entry in instance network info cache for port 7ce08c90-28e1-4989-9616-05bc8d5ee616. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.662823] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [{"id": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "address": "fa:16:3e:c5:e0:b6", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce08c90-28", "ovs_interfaceid": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.670762] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.968535] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:34Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 961.968918] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.969039] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 961.969265] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.969450] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 961.969644] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 961.969867] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 961.970048] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 961.970231] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 961.970403] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 961.970583] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 961.980520] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3a5b033-6a81-4757-a891-7bca70247ef4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.001799] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 962.001799] env[61911]: value = "task-1251427" [ 962.001799] env[61911]: _type = "Task" [ 962.001799] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.009642] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251427, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.057717] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251426, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064148} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.057717] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.058524] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70135a3-dac5-4340-a987-4928fdd37fc7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.081302] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 7060c2a1-7546-41f9-891d-af6a59fa1aa5/7060c2a1-7546-41f9-891d-af6a59fa1aa5.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.085962] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18fc2457-7d57-4bd4-99d6-da20b9b1ab5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.107178] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 962.107178] env[61911]: value = "task-1251428" [ 962.107178] env[61911]: _type = "Task" [ 962.107178] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.118062] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251428, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.165523] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Releasing lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.165718] env[61911]: DEBUG nova.compute.manager [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Received event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 962.165901] env[61911]: DEBUG nova.compute.manager [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing instance network info cache due to event network-changed-7ce08c90-28e1-4989-9616-05bc8d5ee616. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 962.166168] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Acquiring lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.166321] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Acquired lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.166518] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Refreshing network info cache for port 7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.216982] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460687ac-81ba-46c6-b13b-c3762678f64d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.227050] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ad1947-3c08-4748-81ec-3d272093d930 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.233421] env[61911]: DEBUG nova.network.neutron [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Updated VIF entry in instance network info cache for port 33e8e080-830e-4a6a-8e9b-39eded7bf6de. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.233802] env[61911]: DEBUG nova.network.neutron [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Updating instance_info_cache with network_info: [{"id": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "address": "fa:16:3e:a4:2b:ca", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33e8e080-83", "ovs_interfaceid": "33e8e080-830e-4a6a-8e9b-39eded7bf6de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.260206] env[61911]: DEBUG oslo_concurrency.lockutils [req-7e5cbaa3-0ec6-4a05-8634-2788ac2723dd req-f9af09ce-ae83-422d-840f-290310c499af service nova] Releasing lock "refresh_cache-fddba36d-1b15-43fb-9e99-68880e8a235c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.263553] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e204bc8-9a33-4e10-b36d-e2056edc39b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.272151] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9edd6b0-6a34-4baa-8a96-b04dd453ffe5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.286295] env[61911]: DEBUG nova.compute.provider_tree [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.300122] env[61911]: DEBUG nova.network.neutron [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [{"id": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "address": "fa:16:3e:c0:e6:50", "network": {"id": "e9fd250b-3177-4bd7-b71c-d11693f894c3", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-719590582-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e00325b35c074e92a5c80bf2a2df08dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6d1cfde-67", "ovs_interfaceid": "f6d1cfde-6743-4e5f-ac24-8b71f63c293c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.511559] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251427, 'name': ReconfigVM_Task, 'duration_secs': 0.45256} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.511911] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance 'f386424d-953d-431d-a56b-8542f2a3458c' progress to 33 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.602759] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "3831af9a-fe2a-49d6-9e38-fc78e2616461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.602967] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.617361] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251428, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.789309] env[61911]: DEBUG nova.scheduler.client.report [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 962.802484] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Releasing lock "refresh_cache-8051e8ec-6d82-416c-858a-f0cdd00f869d" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.803450] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bdb098-3c70-419f-ba30-7caa941ed4ca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.809692] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Resuming the VM {{(pid=61911) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 962.810527] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a252087-0c99-4ef2-af56-d7b949b75e11 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.819324] env[61911]: DEBUG oslo_vmware.api [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 962.819324] env[61911]: value = "task-1251429" [ 962.819324] env[61911]: _type = "Task" [ 962.819324] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.826718] env[61911]: DEBUG oslo_vmware.api [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.880760] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updated VIF entry in instance network info cache for port 7ce08c90-28e1-4989-9616-05bc8d5ee616. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.881172] env[61911]: DEBUG nova.network.neutron [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [{"id": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "address": "fa:16:3e:c5:e0:b6", "network": {"id": "7b388762-1e85-4a70-b5af-626ddd1b9c89", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1257418449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f9c20b59cfc4a8389da83b0e6a87730", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4225eb1f-0af4-4ed4-8e3d-de822eb6d4ea", "external-id": "nsx-vlan-transportzone-40", "segmentation_id": 40, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce08c90-28", "ovs_interfaceid": "7ce08c90-28e1-4989-9616-05bc8d5ee616", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.019044] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 963.019472] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.019554] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 963.019680] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.019831] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 963.019981] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 963.020205] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 963.020372] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 963.020566] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 963.020748] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 963.021012] env[61911]: DEBUG nova.virt.hardware [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 963.026436] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Reconfiguring VM instance instance-00000055 to detach disk 2000 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 963.027064] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5be2131f-0780-4266-8632-47f37196b210 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.045294] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 963.045294] env[61911]: value = "task-1251430" [ 963.045294] env[61911]: _type = "Task" [ 963.045294] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.053155] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251430, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.106048] env[61911]: DEBUG nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 963.117055] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251428, 'name': ReconfigVM_Task, 'duration_secs': 0.899744} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.117055] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 7060c2a1-7546-41f9-891d-af6a59fa1aa5/7060c2a1-7546-41f9-891d-af6a59fa1aa5.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.117654] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52ea263c-7f76-4506-b313-f0561ed90bbe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.123755] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 963.123755] env[61911]: value = "task-1251431" [ 963.123755] env[61911]: _type = "Task" [ 963.123755] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.131714] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251431, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.294757] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.295488] env[61911]: DEBUG nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 963.298070] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.635s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.299796] env[61911]: INFO nova.compute.claims [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.329374] env[61911]: DEBUG oslo_vmware.api [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251429, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.384559] env[61911]: DEBUG oslo_concurrency.lockutils [req-956400da-b692-412c-ab0c-3429aec37961 req-27bef50d-7c60-4f0a-abbe-5ac0a10b7c89 service nova] Releasing lock "refresh_cache-5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.555101] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251430, 'name': ReconfigVM_Task, 'duration_secs': 0.14821} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.555365] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Reconfigured VM instance instance-00000055 to detach disk 2000 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 963.556197] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2aef4e-96c7-4555-b99b-79b2b9511d80 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.580402] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] f386424d-953d-431d-a56b-8542f2a3458c/f386424d-953d-431d-a56b-8542f2a3458c.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.580741] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-061b9e49-b4c0-43a5-bd7e-8b63208f9587 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.599655] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 963.599655] env[61911]: value = "task-1251432" [ 963.599655] env[61911]: _type = "Task" [ 963.599655] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.607849] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251432, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.635729] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251431, 'name': Rename_Task, 'duration_secs': 0.173203} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.636626] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.636920] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39615a0f-c43c-4dd4-b518-55b822a2f37f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.640030] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.642616] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 963.642616] env[61911]: value = "task-1251433" [ 963.642616] env[61911]: _type = "Task" [ 963.642616] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.650788] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.804593] env[61911]: DEBUG nova.compute.utils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.808892] env[61911]: DEBUG nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 963.809231] env[61911]: DEBUG nova.network.neutron [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 963.831787] env[61911]: DEBUG oslo_vmware.api [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251429, 'name': PowerOnVM_Task, 'duration_secs': 0.643815} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.832134] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Resumed the VM {{(pid=61911) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 963.832432] env[61911]: DEBUG nova.compute.manager [None req-0efc11ad-10b0-48f2-b45f-ff64b9cccc1e tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 963.833268] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c92028-f1d1-4b1b-9cbc-a7a37aa1de0e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.861901] env[61911]: DEBUG nova.policy [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b258257c7d84a90913c462d7e33e4be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fe1ad57908647b79849aea9d81726a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 964.105361] env[61911]: DEBUG nova.network.neutron [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Successfully created port: 450ba0e3-64ce-4aff-bf9d-059d8aa237fd {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.112472] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251432, 'name': ReconfigVM_Task, 'duration_secs': 0.284} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.112789] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Reconfigured VM instance instance-00000055 to attach disk [datastore1] f386424d-953d-431d-a56b-8542f2a3458c/f386424d-953d-431d-a56b-8542f2a3458c.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.117019] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance 'f386424d-953d-431d-a56b-8542f2a3458c' progress to 50 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 964.153761] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251433, 'name': PowerOnVM_Task} progress is 79%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.313860] env[61911]: DEBUG nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 964.525683] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732e71ef-7d97-41ba-b514-32524608e80e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.534110] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5fec40-0044-44ab-afdf-412edbb6e8ca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.563906] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf43e52d-bbde-4b25-bb2d-1be76fb97724 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.572052] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27697c04-2b71-45d3-af25-b2d4e62e08f3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.584127] env[61911]: DEBUG nova.compute.provider_tree [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.623014] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932bd211-e0d0-4459-88b9-dec6d6845a05 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.642784] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add621f9-9b1b-4290-8882-974859327e86 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.663969] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance 'f386424d-953d-431d-a56b-8542f2a3458c' progress to 67 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 964.670534] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251433, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.087091] env[61911]: DEBUG nova.scheduler.client.report [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 965.155310] env[61911]: DEBUG oslo_vmware.api [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251433, 'name': PowerOnVM_Task, 'duration_secs': 1.066074} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.155697] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 965.155874] env[61911]: INFO nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Took 8.36 seconds to spawn the instance on the hypervisor. [ 965.156035] env[61911]: DEBUG nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 965.156788] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c469547e-4014-40fe-8f99-284ee9f3565b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.235363] env[61911]: DEBUG nova.network.neutron [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Port 5937b7a1-0418-4e62-9605-d52f38ca02b4 binding to destination host cpu-1 is already ACTIVE {{(pid=61911) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 965.323652] env[61911]: DEBUG nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 965.348591] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 965.348844] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.349042] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 965.349241] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.349392] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 965.349543] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 965.349785] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 965.349956] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 965.350157] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 965.350324] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 965.350510] env[61911]: DEBUG nova.virt.hardware [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 965.351515] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62eba718-41ac-42a6-9efd-6a977cb924ca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.360966] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa478f0-5022-4c09-9517-394ad722e18e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.501836] env[61911]: DEBUG nova.compute.manager [req-78119e6d-2d17-4854-ba38-5a2dea095fc7 req-4e6c7500-7ccc-44be-a7bb-a18a8c73e8c1 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Received event network-vif-plugged-450ba0e3-64ce-4aff-bf9d-059d8aa237fd {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 965.502089] env[61911]: DEBUG oslo_concurrency.lockutils [req-78119e6d-2d17-4854-ba38-5a2dea095fc7 req-4e6c7500-7ccc-44be-a7bb-a18a8c73e8c1 service nova] Acquiring lock "b1d2b205-81d7-410b-901d-ccd608bc5bec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.504578] env[61911]: DEBUG oslo_concurrency.lockutils [req-78119e6d-2d17-4854-ba38-5a2dea095fc7 req-4e6c7500-7ccc-44be-a7bb-a18a8c73e8c1 service nova] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.504578] env[61911]: DEBUG oslo_concurrency.lockutils [req-78119e6d-2d17-4854-ba38-5a2dea095fc7 req-4e6c7500-7ccc-44be-a7bb-a18a8c73e8c1 service nova] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.504578] env[61911]: DEBUG nova.compute.manager [req-78119e6d-2d17-4854-ba38-5a2dea095fc7 req-4e6c7500-7ccc-44be-a7bb-a18a8c73e8c1 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] No waiting events found dispatching network-vif-plugged-450ba0e3-64ce-4aff-bf9d-059d8aa237fd {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 965.504578] env[61911]: WARNING nova.compute.manager [req-78119e6d-2d17-4854-ba38-5a2dea095fc7 req-4e6c7500-7ccc-44be-a7bb-a18a8c73e8c1 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Received unexpected event network-vif-plugged-450ba0e3-64ce-4aff-bf9d-059d8aa237fd for instance with vm_state building and task_state spawning. [ 965.592696] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.593184] env[61911]: DEBUG nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 965.596324] env[61911]: DEBUG nova.network.neutron [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Successfully updated port: 450ba0e3-64ce-4aff-bf9d-059d8aa237fd {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.597694] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.816s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.597919] env[61911]: DEBUG nova.objects.instance [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lazy-loading 'resources' on Instance uuid 3c4c8943-5324-4c3e-b7e0-7ea5972bc026 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.674592] env[61911]: INFO nova.compute.manager [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Took 13.31 seconds to build instance. [ 966.100279] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.100438] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.100592] env[61911]: DEBUG nova.network.neutron [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.103586] env[61911]: DEBUG nova.compute.utils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.107318] env[61911]: DEBUG nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 966.107488] env[61911]: DEBUG nova.network.neutron [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.149430] env[61911]: DEBUG nova.policy [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b258257c7d84a90913c462d7e33e4be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fe1ad57908647b79849aea9d81726a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.177168] env[61911]: DEBUG oslo_concurrency.lockutils [None req-694338d0-9b2f-4b2e-8efa-311d3c40a2b0 tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.822s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.261953] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f386424d-953d-431d-a56b-8542f2a3458c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.262294] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.262487] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.330817] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2a10d9-4cc0-42a9-be8b-dd1798c81b40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.339941] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eaebf91-b42d-44fc-9803-42e2242a8ff1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.371175] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7ca46f-a349-44a5-83d1-1704489733cf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.379324] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85381d1-f74f-49b9-806c-5d7d2b811bb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.396961] env[61911]: DEBUG nova.compute.provider_tree [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.413188] env[61911]: DEBUG nova.network.neutron [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Successfully created port: 022b3630-95b4-40fa-8930-c97b6b008fd3 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 966.612270] env[61911]: DEBUG nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 966.635958] env[61911]: DEBUG nova.network.neutron [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 966.799188] env[61911]: DEBUG nova.network.neutron [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Updating instance_info_cache with network_info: [{"id": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "address": "fa:16:3e:f9:7d:ad", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450ba0e3-64", "ovs_interfaceid": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.897266] env[61911]: DEBUG nova.scheduler.client.report [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 967.297522] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.297886] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.298033] env[61911]: DEBUG nova.network.neutron [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.301150] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.301804] env[61911]: DEBUG nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Instance network_info: |[{"id": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "address": "fa:16:3e:f9:7d:ad", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450ba0e3-64", "ovs_interfaceid": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 967.301982] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:7d:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '450ba0e3-64ce-4aff-bf9d-059d8aa237fd', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.310861] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Creating folder: Project (9fe1ad57908647b79849aea9d81726a8). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 967.311942] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb5f06b0-2dca-4a72-8987-ae470ccc97a7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.322972] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Created folder: Project (9fe1ad57908647b79849aea9d81726a8) in parent group-v269521. [ 967.323236] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Creating folder: Instances. Parent ref: group-v269646. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 967.323519] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb8f3477-be6f-440a-9dda-9bbb2872be44 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.333553] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Created folder: Instances in parent group-v269646. [ 967.333849] env[61911]: DEBUG oslo.service.loopingcall [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.334087] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.334328] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82ef2c80-13cf-4cfe-9edb-4b45606b5b35 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.354206] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.354206] env[61911]: value = "task-1251436" [ 967.354206] env[61911]: _type = "Task" [ 967.354206] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.362155] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251436, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.391565] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.391819] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.392047] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.392245] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.392426] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.395079] env[61911]: INFO nova.compute.manager [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Terminating instance [ 967.403399] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.805s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.405972] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.080s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.407675] env[61911]: INFO nova.compute.claims [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.432034] env[61911]: INFO nova.scheduler.client.report [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Deleted allocations for instance 3c4c8943-5324-4c3e-b7e0-7ea5972bc026 [ 967.533031] env[61911]: DEBUG nova.compute.manager [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Received event network-changed-450ba0e3-64ce-4aff-bf9d-059d8aa237fd {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 967.533339] env[61911]: DEBUG nova.compute.manager [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Refreshing instance network info cache due to event network-changed-450ba0e3-64ce-4aff-bf9d-059d8aa237fd. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 967.533639] env[61911]: DEBUG oslo_concurrency.lockutils [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] Acquiring lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.533846] env[61911]: DEBUG oslo_concurrency.lockutils [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] Acquired lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.534072] env[61911]: DEBUG nova.network.neutron [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Refreshing network info cache for port 450ba0e3-64ce-4aff-bf9d-059d8aa237fd {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.624375] env[61911]: DEBUG nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 967.651297] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 967.651571] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.651741] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 967.651930] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.652103] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 967.652262] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 967.652482] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 967.652648] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 967.652830] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 967.653057] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 967.653252] env[61911]: DEBUG nova.virt.hardware [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 967.654153] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3239fd26-0678-4e2f-9f3e-f832bebe4f2e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.662612] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1106b8-8689-4797-877e-8b86040594a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.868062] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251436, 'name': CreateVM_Task, 'duration_secs': 0.388366} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.871822] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 967.872880] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.873275] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.873848] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.874654] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f509130e-4867-4a6d-a9ab-f691891440d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.880791] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 967.880791] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280b4de-dfda-4138-8056-f59a6c702b7d" [ 967.880791] env[61911]: _type = "Task" [ 967.880791] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.891686] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280b4de-dfda-4138-8056-f59a6c702b7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.899903] env[61911]: DEBUG nova.compute.manager [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 967.900120] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.901008] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37963197-a04a-48ca-ad5e-1d631b74d39f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.908136] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.908585] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-057dc255-9c15-4e63-aea6-654748bad8be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.915930] env[61911]: DEBUG oslo_vmware.api [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 967.915930] env[61911]: value = "task-1251437" [ 967.915930] env[61911]: _type = "Task" [ 967.915930] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.926873] env[61911]: DEBUG oslo_vmware.api [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.927461] env[61911]: DEBUG nova.network.neutron [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Successfully updated port: 022b3630-95b4-40fa-8930-c97b6b008fd3 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.939012] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e739c2a0-cb05-40f7-9456-84376965a9af tempest-ImagesTestJSON-2014078941 tempest-ImagesTestJSON-2014078941-project-member] Lock "3c4c8943-5324-4c3e-b7e0-7ea5972bc026" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.330s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.035373] env[61911]: DEBUG nova.network.neutron [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance_info_cache with network_info: [{"id": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "address": "fa:16:3e:ef:fb:2e", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5937b7a1-04", "ovs_interfaceid": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.396937] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280b4de-dfda-4138-8056-f59a6c702b7d, 'name': SearchDatastore_Task, 'duration_secs': 0.010131} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.398458] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.398968] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.399486] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.399828] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.400662] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.401126] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98f7a271-ece6-4bcf-b144-171b7a6d005e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.411292] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.411525] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.412367] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51baa456-8e3b-4f30-85cd-9dcc9b4ea8d1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.421088] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 968.421088] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525715d6-4263-a284-9bac-aba9a4b878f0" [ 968.421088] env[61911]: _type = "Task" [ 968.421088] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.429870] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "refresh_cache-b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.430105] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "refresh_cache-b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.430342] env[61911]: DEBUG nova.network.neutron [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.442618] env[61911]: DEBUG oslo_vmware.api [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251437, 'name': PowerOffVM_Task, 'duration_secs': 0.215847} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.446333] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.446528] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.447042] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525715d6-4263-a284-9bac-aba9a4b878f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.449909] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdb3d551-b41a-451a-a628-f744219a5939 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.528708] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.528857] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.529057] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Deleting the datastore file [datastore2] 7060c2a1-7546-41f9-891d-af6a59fa1aa5 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.532100] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05181921-0eab-46ce-9b0e-4dcb58ae9465 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.539165] env[61911]: DEBUG oslo_vmware.api [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for the task: (returnval){ [ 968.539165] env[61911]: value = "task-1251439" [ 968.539165] env[61911]: _type = "Task" [ 968.539165] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.543338] env[61911]: DEBUG oslo_concurrency.lockutils [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.555533] env[61911]: DEBUG oslo_vmware.api [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251439, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.651726] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5b8f80-65c4-43a7-ae46-f79572baa9bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.661534] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1407a46-973e-4005-8eca-b17c462d2aed {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.699633] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.699909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.700245] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.700450] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.700631] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.707283] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78afcdee-7c61-41e6-84a1-c085e756de69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.711391] env[61911]: INFO nova.compute.manager [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Terminating instance [ 968.722308] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4a99fb-29eb-45c8-a2e2-13fdaf632d4c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.736853] env[61911]: DEBUG nova.compute.provider_tree [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.765211] env[61911]: DEBUG nova.network.neutron [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Updated VIF entry in instance network info cache for port 450ba0e3-64ce-4aff-bf9d-059d8aa237fd. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.765594] env[61911]: DEBUG nova.network.neutron [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Updating instance_info_cache with network_info: [{"id": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "address": "fa:16:3e:f9:7d:ad", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450ba0e3-64", "ovs_interfaceid": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.942565] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525715d6-4263-a284-9bac-aba9a4b878f0, 'name': SearchDatastore_Task, 'duration_secs': 0.022814} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.943750] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38520cd2-2c81-4869-8a38-ec2081baed66 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.948945] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 968.948945] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ba0142-486a-df0d-e740-4df19da25933" [ 968.948945] env[61911]: _type = "Task" [ 968.948945] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.957702] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ba0142-486a-df0d-e740-4df19da25933, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.979636] env[61911]: DEBUG nova.network.neutron [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 969.049789] env[61911]: DEBUG oslo_vmware.api [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Task: {'id': task-1251439, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325985} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.050285] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.050517] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.050700] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.050877] env[61911]: INFO nova.compute.manager [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 969.051151] env[61911]: DEBUG oslo.service.loopingcall [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.056694] env[61911]: DEBUG nova.compute.manager [-] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 969.056790] env[61911]: DEBUG nova.network.neutron [-] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 969.072745] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454df060-05a9-4d4b-a124-fbbbaabdacbd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.093535] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4dba000-048c-43b7-b159-dd226d0c1027 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.102844] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance 'f386424d-953d-431d-a56b-8542f2a3458c' progress to 83 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 969.140582] env[61911]: DEBUG nova.network.neutron [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Updating instance_info_cache with network_info: [{"id": "022b3630-95b4-40fa-8930-c97b6b008fd3", "address": "fa:16:3e:a0:73:ef", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022b3630-95", "ovs_interfaceid": "022b3630-95b4-40fa-8930-c97b6b008fd3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.216064] env[61911]: DEBUG nova.compute.manager [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 969.216331] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.217665] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414c3b97-68f1-4c04-a43c-87354f63d417 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.225866] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.226168] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60a4e4d5-81b7-4129-9f3c-3be8d48f8d04 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.233250] env[61911]: DEBUG oslo_vmware.api [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 969.233250] env[61911]: value = "task-1251440" [ 969.233250] env[61911]: _type = "Task" [ 969.233250] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.243503] env[61911]: DEBUG nova.scheduler.client.report [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 969.246875] env[61911]: DEBUG oslo_vmware.api [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251440, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.268217] env[61911]: DEBUG oslo_concurrency.lockutils [req-b026c4f3-0830-4648-b0a8-c122d8d7de17 req-582e16a8-b09d-4cb3-8961-3f432de46519 service nova] Releasing lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.459733] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ba0142-486a-df0d-e740-4df19da25933, 'name': SearchDatastore_Task, 'duration_secs': 0.025151} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.460036] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.460320] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] b1d2b205-81d7-410b-901d-ccd608bc5bec/b1d2b205-81d7-410b-901d-ccd608bc5bec.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.460587] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a09f962-033e-494e-89ba-8adce3274c56 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.466724] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 969.466724] env[61911]: value = "task-1251441" [ 969.466724] env[61911]: _type = "Task" [ 969.466724] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.474484] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251441, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.572829] env[61911]: DEBUG nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Received event network-vif-plugged-022b3630-95b4-40fa-8930-c97b6b008fd3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 969.572829] env[61911]: DEBUG oslo_concurrency.lockutils [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] Acquiring lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.572829] env[61911]: DEBUG oslo_concurrency.lockutils [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.572829] env[61911]: DEBUG oslo_concurrency.lockutils [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.573023] env[61911]: DEBUG nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] No waiting events found dispatching network-vif-plugged-022b3630-95b4-40fa-8930-c97b6b008fd3 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 969.573221] env[61911]: WARNING nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Received unexpected event network-vif-plugged-022b3630-95b4-40fa-8930-c97b6b008fd3 for instance with vm_state building and task_state spawning. [ 969.573449] env[61911]: DEBUG nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Received event network-changed-022b3630-95b4-40fa-8930-c97b6b008fd3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 969.573629] env[61911]: DEBUG nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Refreshing instance network info cache due to event network-changed-022b3630-95b4-40fa-8930-c97b6b008fd3. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 969.573803] env[61911]: DEBUG oslo_concurrency.lockutils [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] Acquiring lock "refresh_cache-b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.613713] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.614136] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c064629-a3dd-419e-abbd-ef59ce9c6934 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.622946] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 969.622946] env[61911]: value = "task-1251442" [ 969.622946] env[61911]: _type = "Task" [ 969.622946] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.630894] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251442, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.644031] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "refresh_cache-b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.644031] env[61911]: DEBUG nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Instance network_info: |[{"id": "022b3630-95b4-40fa-8930-c97b6b008fd3", "address": "fa:16:3e:a0:73:ef", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022b3630-95", "ovs_interfaceid": "022b3630-95b4-40fa-8930-c97b6b008fd3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 969.644807] env[61911]: DEBUG oslo_concurrency.lockutils [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] Acquired lock "refresh_cache-b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.645017] env[61911]: DEBUG nova.network.neutron [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Refreshing network info cache for port 022b3630-95b4-40fa-8930-c97b6b008fd3 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 969.646713] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:73:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '022b3630-95b4-40fa-8930-c97b6b008fd3', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.657712] env[61911]: DEBUG oslo.service.loopingcall [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.659215] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.659215] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b116600a-ebe1-4b74-a079-9930ad5f122d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.682213] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.682213] env[61911]: value = "task-1251443" [ 969.682213] env[61911]: _type = "Task" [ 969.682213] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.691558] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251443, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.743627] env[61911]: DEBUG oslo_vmware.api [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251440, 'name': PowerOffVM_Task, 'duration_secs': 0.212265} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.744033] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.744249] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.744569] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2bb28c0-9137-4324-b59c-9a644f3d66b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.747632] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.748127] env[61911]: DEBUG nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 969.750684] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.080s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.750914] env[61911]: DEBUG nova.objects.instance [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'resources' on Instance uuid ec53a8ee-df42-4003-a933-17aa4c90c3aa {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.851863] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.852115] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.852410] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleting the datastore file [datastore1] 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.852722] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d54b737-1685-4bf8-b7c0-02a2decca327 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.861012] env[61911]: DEBUG oslo_vmware.api [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for the task: (returnval){ [ 969.861012] env[61911]: value = "task-1251445" [ 969.861012] env[61911]: _type = "Task" [ 969.861012] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.870655] env[61911]: DEBUG oslo_vmware.api [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.978457] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251441, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447227} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.978985] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] b1d2b205-81d7-410b-901d-ccd608bc5bec/b1d2b205-81d7-410b-901d-ccd608bc5bec.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 969.979462] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.980135] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf4e3195-cfb4-4481-b844-82738fa269a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.986977] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 969.986977] env[61911]: value = "task-1251446" [ 969.986977] env[61911]: _type = "Task" [ 969.986977] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.996658] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251446, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.009595] env[61911]: DEBUG nova.network.neutron [-] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.133872] env[61911]: DEBUG oslo_vmware.api [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251442, 'name': PowerOnVM_Task, 'duration_secs': 0.469552} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.134470] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.137091] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b3b3e3-39d6-4962-80c3-5156bf780921 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance 'f386424d-953d-431d-a56b-8542f2a3458c' progress to 100 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 970.191751] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251443, 'name': CreateVM_Task, 'duration_secs': 0.497131} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.194080] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.194934] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.195125] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.195449] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 970.195987] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25ee653a-dc6d-4c67-bc13-c16cf044b334 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.200456] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 970.200456] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a7b089-16d6-89bf-99c8-5d2190939a9a" [ 970.200456] env[61911]: _type = "Task" [ 970.200456] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.208700] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a7b089-16d6-89bf-99c8-5d2190939a9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.253968] env[61911]: DEBUG nova.compute.utils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.257800] env[61911]: DEBUG nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 970.258029] env[61911]: DEBUG nova.network.neutron [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.298069] env[61911]: DEBUG nova.policy [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b258257c7d84a90913c462d7e33e4be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fe1ad57908647b79849aea9d81726a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.374174] env[61911]: DEBUG oslo_vmware.api [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Task: {'id': task-1251445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188869} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.377229] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.377479] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.377688] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.377931] env[61911]: INFO nova.compute.manager [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 970.378156] env[61911]: DEBUG oslo.service.loopingcall [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.378572] env[61911]: DEBUG nova.compute.manager [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 970.378674] env[61911]: DEBUG nova.network.neutron [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.460962] env[61911]: DEBUG nova.network.neutron [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Updated VIF entry in instance network info cache for port 022b3630-95b4-40fa-8930-c97b6b008fd3. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.461677] env[61911]: DEBUG nova.network.neutron [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Updating instance_info_cache with network_info: [{"id": "022b3630-95b4-40fa-8930-c97b6b008fd3", "address": "fa:16:3e:a0:73:ef", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap022b3630-95", "ovs_interfaceid": "022b3630-95b4-40fa-8930-c97b6b008fd3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.495785] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16a40dc-e258-46a7-bd88-46bf1b643afa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.501622] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251446, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075837} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.502640] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.503290] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f0c434-54cb-46df-809f-ceacc444aaa6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.508433] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b13f59-e65e-4c67-ba8b-7a901056cb09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.521084] env[61911]: INFO nova.compute.manager [-] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Took 1.46 seconds to deallocate network for instance. [ 970.530179] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] b1d2b205-81d7-410b-901d-ccd608bc5bec/b1d2b205-81d7-410b-901d-ccd608bc5bec.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.532670] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bff6be7-4606-4291-84a2-43ec2b25d7cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.579650] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca4038c-cc25-4889-8e2d-bd48daeeba27 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.582834] env[61911]: DEBUG nova.network.neutron [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Successfully created port: 227cefa9-0b37-43fc-9c0b-da58e16ae969 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.586420] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 970.586420] env[61911]: value = "task-1251447" [ 970.586420] env[61911]: _type = "Task" [ 970.586420] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.592965] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f7779d-7449-49dc-a547-c2028cc4732f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.600218] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251447, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.611584] env[61911]: DEBUG nova.compute.provider_tree [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.710922] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a7b089-16d6-89bf-99c8-5d2190939a9a, 'name': SearchDatastore_Task, 'duration_secs': 0.042323} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.711664] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.712038] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.712299] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.712509] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.712750] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.713089] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76e362be-5a39-4913-ac88-5987950426ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.725242] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.725390] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.727609] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d079bdea-46d1-4d16-b9e5-4b488d651230 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.732616] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 970.732616] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527675ba-555a-757f-54fe-741f569738e6" [ 970.732616] env[61911]: _type = "Task" [ 970.732616] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.740357] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527675ba-555a-757f-54fe-741f569738e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.762028] env[61911]: DEBUG nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 970.967525] env[61911]: DEBUG oslo_concurrency.lockutils [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] Releasing lock "refresh_cache-b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.967687] env[61911]: DEBUG nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Received event network-vif-deleted-a3f9143d-1bd7-42d4-be27-f2a515d78820 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 970.968022] env[61911]: INFO nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Neutron deleted interface a3f9143d-1bd7-42d4-be27-f2a515d78820; detaching it from the instance and deleting it from the info cache [ 970.968223] env[61911]: DEBUG nova.network.neutron [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.054419] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.097846] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251447, 'name': ReconfigVM_Task, 'duration_secs': 0.299052} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.098204] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Reconfigured VM instance instance-00000059 to attach disk [datastore1] b1d2b205-81d7-410b-901d-ccd608bc5bec/b1d2b205-81d7-410b-901d-ccd608bc5bec.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.098850] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2659577-8864-45f5-9d3b-77feba528957 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.105802] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 971.105802] env[61911]: value = "task-1251448" [ 971.105802] env[61911]: _type = "Task" [ 971.105802] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.114220] env[61911]: DEBUG nova.scheduler.client.report [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 971.117225] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251448, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.172122] env[61911]: DEBUG nova.network.neutron [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.241801] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527675ba-555a-757f-54fe-741f569738e6, 'name': SearchDatastore_Task, 'duration_secs': 0.008248} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.243036] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f67c1efc-2c5a-4e78-acb5-4ee00eba4f22 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.247802] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 971.247802] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c49747-1913-811b-dfc2-40a31049da75" [ 971.247802] env[61911]: _type = "Task" [ 971.247802] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.255399] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c49747-1913-811b-dfc2-40a31049da75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.472142] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2fbaec45-326b-4b3d-947f-9a9d89c85285 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.481793] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58715ad-5e21-4772-b73f-02ec13a783cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.509349] env[61911]: DEBUG nova.compute.manager [req-378e7441-059b-41dc-9062-ae2bdd40b4ba req-db80e01a-da5c-426c-8d50-a12b40093281 service nova] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Detach interface failed, port_id=a3f9143d-1bd7-42d4-be27-f2a515d78820, reason: Instance 7060c2a1-7546-41f9-891d-af6a59fa1aa5 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 971.595737] env[61911]: DEBUG nova.compute.manager [req-681c4d89-369f-4ff0-92df-4627b140369d req-34bef5be-cf1d-454c-b53b-df1c05700c01 service nova] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Received event network-vif-deleted-f6d1cfde-6743-4e5f-ac24-8b71f63c293c {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 971.615225] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251448, 'name': Rename_Task, 'duration_secs': 0.147075} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.615472] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 971.615699] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30574d4a-7e34-43db-93a5-a833d4cc8298 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.618959] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.868s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.622090] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.982s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.623468] env[61911]: INFO nova.compute.claims [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.625755] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 971.625755] env[61911]: value = "task-1251449" [ 971.625755] env[61911]: _type = "Task" [ 971.625755] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.633374] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.643034] env[61911]: INFO nova.scheduler.client.report [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted allocations for instance ec53a8ee-df42-4003-a933-17aa4c90c3aa [ 971.675029] env[61911]: INFO nova.compute.manager [-] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Took 1.30 seconds to deallocate network for instance. [ 971.758944] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c49747-1913-811b-dfc2-40a31049da75, 'name': SearchDatastore_Task, 'duration_secs': 0.008672} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.759229] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.759539] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8/b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 971.759770] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da997d5f-be12-4c11-9e83-f4cf136fab4d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.766451] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 971.766451] env[61911]: value = "task-1251450" [ 971.766451] env[61911]: _type = "Task" [ 971.766451] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.769993] env[61911]: DEBUG nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 971.776772] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251450, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.794826] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:34Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 971.795115] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.795286] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 971.795477] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.795631] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 971.795811] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 971.796079] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 971.796256] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 971.796436] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 971.796608] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 971.796785] env[61911]: DEBUG nova.virt.hardware [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 971.797720] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a105d4fc-4252-46e2-a16c-5d184b7ce90c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.805382] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187f568e-1fff-4620-a997-4a04ab489306 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.020030] env[61911]: DEBUG nova.compute.manager [req-b2541015-4196-48b5-894a-14373120700b req-e1e046d8-27b9-4179-a549-e094273a39d7 service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Received event network-vif-plugged-227cefa9-0b37-43fc-9c0b-da58e16ae969 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 972.020030] env[61911]: DEBUG oslo_concurrency.lockutils [req-b2541015-4196-48b5-894a-14373120700b req-e1e046d8-27b9-4179-a549-e094273a39d7 service nova] Acquiring lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.023059] env[61911]: DEBUG oslo_concurrency.lockutils [req-b2541015-4196-48b5-894a-14373120700b req-e1e046d8-27b9-4179-a549-e094273a39d7 service nova] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.023432] env[61911]: DEBUG oslo_concurrency.lockutils [req-b2541015-4196-48b5-894a-14373120700b req-e1e046d8-27b9-4179-a549-e094273a39d7 service nova] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.003s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.023716] env[61911]: DEBUG nova.compute.manager [req-b2541015-4196-48b5-894a-14373120700b req-e1e046d8-27b9-4179-a549-e094273a39d7 service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] No waiting events found dispatching network-vif-plugged-227cefa9-0b37-43fc-9c0b-da58e16ae969 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 972.023982] env[61911]: WARNING nova.compute.manager [req-b2541015-4196-48b5-894a-14373120700b req-e1e046d8-27b9-4179-a549-e094273a39d7 service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Received unexpected event network-vif-plugged-227cefa9-0b37-43fc-9c0b-da58e16ae969 for instance with vm_state building and task_state spawning. [ 972.146815] env[61911]: DEBUG oslo_vmware.api [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251449, 'name': PowerOnVM_Task, 'duration_secs': 0.47266} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.146815] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 972.146815] env[61911]: INFO nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Took 6.82 seconds to spawn the instance on the hypervisor. [ 972.146815] env[61911]: DEBUG nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 972.146815] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb0ab34-b892-48e2-b747-4e1b438d3398 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.156219] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4a6b0a50-9a07-4499-a350-efc6773efd7e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "ec53a8ee-df42-4003-a933-17aa4c90c3aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.065s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.179760] env[61911]: DEBUG nova.network.neutron [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Successfully updated port: 227cefa9-0b37-43fc-9c0b-da58e16ae969 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.184247] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.280568] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251450, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460083} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.283016] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8/b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 972.283016] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.283016] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d4f520d-863c-4532-8677-0bdaa59b96ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.291025] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 972.291025] env[61911]: value = "task-1251451" [ 972.291025] env[61911]: _type = "Task" [ 972.291025] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.299320] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251451, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.677675] env[61911]: INFO nova.compute.manager [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Took 14.78 seconds to build instance. [ 972.678900] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "refresh_cache-d9234a5f-4d26-4bdc-8f32-3120830a8abe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.679386] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "refresh_cache-d9234a5f-4d26-4bdc-8f32-3120830a8abe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.679386] env[61911]: DEBUG nova.network.neutron [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.804030] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251451, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060133} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.804328] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.805150] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b628ccff-c74b-404f-b109-a7c90692dc04 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.829856] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8/b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.832423] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e4e7ea9-3e0c-44c3-adc0-0696699bec96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.853150] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 972.853150] env[61911]: value = "task-1251452" [ 972.853150] env[61911]: _type = "Task" [ 972.853150] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.857522] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129e7b54-1dc6-4fa0-9557-7f6caaa664a5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.865783] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251452, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.868294] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784cbbe7-ddc8-4fe8-8bad-16c78c409b48 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.896979] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859f5abf-3fda-45b6-b6cd-ced8d356e80a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.904033] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19d9f0c-9286-493f-ae1a-d8730d2f91a2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.916446] env[61911]: DEBUG nova.compute.provider_tree [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.181717] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f58a4e2e-b942-4892-84d9-97ff081ed671 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.287s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.186555] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "f63fa31b-8840-4f95-9514-51f7d5393b25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.186792] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.189688] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f386424d-953d-431d-a56b-8542f2a3458c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.189903] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.190102] env[61911]: DEBUG nova.compute.manager [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Going to confirm migration 2 {{(pid=61911) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 973.222148] env[61911]: DEBUG nova.network.neutron [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.358239] env[61911]: DEBUG nova.network.neutron [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Updating instance_info_cache with network_info: [{"id": "227cefa9-0b37-43fc-9c0b-da58e16ae969", "address": "fa:16:3e:8f:da:2b", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227cefa9-0b", "ovs_interfaceid": "227cefa9-0b37-43fc-9c0b-da58e16ae969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.365200] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.420251] env[61911]: DEBUG nova.scheduler.client.report [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 973.688733] env[61911]: DEBUG nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 973.724896] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.725169] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.725397] env[61911]: DEBUG nova.network.neutron [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 973.725635] env[61911]: DEBUG nova.objects.instance [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lazy-loading 'info_cache' on Instance uuid f386424d-953d-431d-a56b-8542f2a3458c {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.863673] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "refresh_cache-d9234a5f-4d26-4bdc-8f32-3120830a8abe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.863930] env[61911]: DEBUG nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Instance network_info: |[{"id": "227cefa9-0b37-43fc-9c0b-da58e16ae969", "address": "fa:16:3e:8f:da:2b", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227cefa9-0b", "ovs_interfaceid": "227cefa9-0b37-43fc-9c0b-da58e16ae969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 973.864266] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251452, 'name': ReconfigVM_Task, 'duration_secs': 0.729791} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.864612] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:da:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '227cefa9-0b37-43fc-9c0b-da58e16ae969', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.871909] env[61911]: DEBUG oslo.service.loopingcall [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.872186] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Reconfigured VM instance instance-0000005a to attach disk [datastore1] b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8/b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.873105] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 973.873317] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99645da8-b2ff-4346-81a9-72f10e239d5d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.874972] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6afe2b6-1c24-4448-84d1-1043c18d6ac7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.893570] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 973.893570] env[61911]: value = "task-1251453" [ 973.893570] env[61911]: _type = "Task" [ 973.893570] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.894753] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 973.894753] env[61911]: value = "task-1251454" [ 973.894753] env[61911]: _type = "Task" [ 973.894753] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.904624] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251454, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.907421] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251453, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.925471] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.926023] env[61911]: DEBUG nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 973.928711] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.874s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.928972] env[61911]: DEBUG nova.objects.instance [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lazy-loading 'resources' on Instance uuid 7060c2a1-7546-41f9-891d-af6a59fa1aa5 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.046338] env[61911]: DEBUG nova.compute.manager [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Received event network-changed-227cefa9-0b37-43fc-9c0b-da58e16ae969 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 974.046545] env[61911]: DEBUG nova.compute.manager [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Refreshing instance network info cache due to event network-changed-227cefa9-0b37-43fc-9c0b-da58e16ae969. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 974.046775] env[61911]: DEBUG oslo_concurrency.lockutils [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] Acquiring lock "refresh_cache-d9234a5f-4d26-4bdc-8f32-3120830a8abe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.046928] env[61911]: DEBUG oslo_concurrency.lockutils [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] Acquired lock "refresh_cache-d9234a5f-4d26-4bdc-8f32-3120830a8abe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.047276] env[61911]: DEBUG nova.network.neutron [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Refreshing network info cache for port 227cefa9-0b37-43fc-9c0b-da58e16ae969 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 974.212325] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.410361] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251454, 'name': CreateVM_Task, 'duration_secs': 0.29409} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.414372] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.414840] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251453, 'name': Rename_Task, 'duration_secs': 0.186565} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.415733] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.416038] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.416540] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.416954] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.417322] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dec9dac-faa1-465b-8b3c-d6b2436c246e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.419638] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b775f2c-55aa-48fc-a1e8-f6614a0196f4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.424791] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 974.424791] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52745f26-6107-c631-36fd-2f63d194611e" [ 974.424791] env[61911]: _type = "Task" [ 974.424791] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.432689] env[61911]: DEBUG nova.compute.utils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.438762] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 974.438762] env[61911]: value = "task-1251455" [ 974.438762] env[61911]: _type = "Task" [ 974.438762] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.439857] env[61911]: DEBUG nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 974.439857] env[61911]: DEBUG nova.network.neutron [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 974.451626] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52745f26-6107-c631-36fd-2f63d194611e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.458097] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251455, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.505439] env[61911]: DEBUG nova.policy [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe50c877fcef489cb798fafe73caaf8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813b3181120245e594809d096a621675', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.662688] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b66c299-ef7c-44d1-a815-616408d65076 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.672184] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3cf51a-349a-4a70-bee2-c77687abac8a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.705166] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b22fa33-a18e-4a5b-baad-fca4c95063ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.713513] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b402c621-6549-4397-b076-ee4e051a1687 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.727617] env[61911]: DEBUG nova.compute.provider_tree [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.823985] env[61911]: DEBUG nova.network.neutron [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Updated VIF entry in instance network info cache for port 227cefa9-0b37-43fc-9c0b-da58e16ae969. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.824398] env[61911]: DEBUG nova.network.neutron [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Updating instance_info_cache with network_info: [{"id": "227cefa9-0b37-43fc-9c0b-da58e16ae969", "address": "fa:16:3e:8f:da:2b", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227cefa9-0b", "ovs_interfaceid": "227cefa9-0b37-43fc-9c0b-da58e16ae969", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.872344] env[61911]: DEBUG nova.network.neutron [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Successfully created port: 4863de42-7e22-4a3c-9024-c0d61f74e3f7 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.939086] env[61911]: DEBUG nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 974.941928] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52745f26-6107-c631-36fd-2f63d194611e, 'name': SearchDatastore_Task, 'duration_secs': 0.018731} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.942561] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.942792] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.943039] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.943196] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.943380] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.948692] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6300bd2d-7a16-4fb7-b5df-3b6ac2432926 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.956533] env[61911]: DEBUG oslo_vmware.api [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251455, 'name': PowerOnVM_Task, 'duration_secs': 0.499236} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.958163] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 974.958390] env[61911]: INFO nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Took 7.33 seconds to spawn the instance on the hypervisor. [ 974.958580] env[61911]: DEBUG nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 974.958882] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.959112] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.960781] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d6dc5a-4261-4a15-92e0-8235e7c390c6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.962805] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4a1b110-1583-4536-8786-2da1d8d8d3da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.968320] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 974.968320] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52efdbcb-cbf3-323c-b3b3-96cdce695c5e" [ 974.968320] env[61911]: _type = "Task" [ 974.968320] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.981910] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52efdbcb-cbf3-323c-b3b3-96cdce695c5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.992172] env[61911]: DEBUG nova.network.neutron [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance_info_cache with network_info: [{"id": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "address": "fa:16:3e:ef:fb:2e", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5937b7a1-04", "ovs_interfaceid": "5937b7a1-0418-4e62-9605-d52f38ca02b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.230973] env[61911]: DEBUG nova.scheduler.client.report [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 975.330394] env[61911]: DEBUG oslo_concurrency.lockutils [req-eb61073c-86f4-4ab3-b84d-899b4aa6eadb req-7e7a79dd-5941-4c99-a283-d1f0cc78393b service nova] Releasing lock "refresh_cache-d9234a5f-4d26-4bdc-8f32-3120830a8abe" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.485292] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52efdbcb-cbf3-323c-b3b3-96cdce695c5e, 'name': SearchDatastore_Task, 'duration_secs': 0.013878} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.488372] env[61911]: INFO nova.compute.manager [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Took 16.84 seconds to build instance. [ 975.489225] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31965b3b-1746-4eaa-bd44-786741e96a76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.494914] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-f386424d-953d-431d-a56b-8542f2a3458c" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.495176] env[61911]: DEBUG nova.objects.instance [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lazy-loading 'migration_context' on Instance uuid f386424d-953d-431d-a56b-8542f2a3458c {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.496314] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 975.496314] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c6cd24-4b4d-da30-a6f6-05e936d5e522" [ 975.496314] env[61911]: _type = "Task" [ 975.496314] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.504130] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c6cd24-4b4d-da30-a6f6-05e936d5e522, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.735767] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.738233] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.554s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.738484] env[61911]: DEBUG nova.objects.instance [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lazy-loading 'resources' on Instance uuid 8051e8ec-6d82-416c-858a-f0cdd00f869d {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.758402] env[61911]: INFO nova.scheduler.client.report [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Deleted allocations for instance 7060c2a1-7546-41f9-891d-af6a59fa1aa5 [ 975.948329] env[61911]: DEBUG nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 975.972783] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 975.973045] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.973216] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 975.973408] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.973562] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 975.973715] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 975.974039] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 975.974151] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 975.974328] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 975.974498] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 975.974676] env[61911]: DEBUG nova.virt.hardware [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 975.975561] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98cdfca-d418-4f4e-be9c-7e7a3336742b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.984078] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea216a5-5806-461e-a708-6d36459cd05d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.997729] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4d7b77a5-8084-41ad-bf87-c340d2ccc0d2 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.360s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.002022] env[61911]: DEBUG nova.objects.base [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 976.002990] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1c265e-73ed-4851-b151-a63ea8b5ec96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.012059] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c6cd24-4b4d-da30-a6f6-05e936d5e522, 'name': SearchDatastore_Task, 'duration_secs': 0.01846} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.024885] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.025188] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] d9234a5f-4d26-4bdc-8f32-3120830a8abe/d9234a5f-4d26-4bdc-8f32-3120830a8abe.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 976.026022] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c306815f-c28b-4988-8943-313a3543ea63 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.027903] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6c599a6-8ea1-4317-84a8-8b9148b5ea3d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.033572] env[61911]: DEBUG oslo_vmware.api [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 976.033572] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f79b66-d42b-86ba-eef2-f27a6181d4d7" [ 976.033572] env[61911]: _type = "Task" [ 976.033572] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.037521] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 976.037521] env[61911]: value = "task-1251456" [ 976.037521] env[61911]: _type = "Task" [ 976.037521] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.044198] env[61911]: DEBUG oslo_vmware.api [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f79b66-d42b-86ba-eef2-f27a6181d4d7, 'name': SearchDatastore_Task, 'duration_secs': 0.007612} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.044841] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.048812] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.268477] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2804d793-f090-4f5a-8a21-e29c2c8c48ae tempest-ServerMetadataTestJSON-1872231897 tempest-ServerMetadataTestJSON-1872231897-project-member] Lock "7060c2a1-7546-41f9-891d-af6a59fa1aa5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.876s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.507503] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe611c5-1cf2-4e1c-b002-10b02dbe95b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.517327] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763d5be0-4ab5-4c52-8f64-8339822eeb07 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.557515] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f72651-5dcc-4799-a08b-65c5cd310ef9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.568913] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53008} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.570277] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f01f22c-08bf-48fc-9584-8ecf3b2822d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.574505] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] d9234a5f-4d26-4bdc-8f32-3120830a8abe/d9234a5f-4d26-4bdc-8f32-3120830a8abe.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.574753] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.575432] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-700c7ed7-a759-44f6-9c90-dc9af2871fec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.588592] env[61911]: DEBUG nova.compute.provider_tree [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.591041] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 976.591041] env[61911]: value = "task-1251457" [ 976.591041] env[61911]: _type = "Task" [ 976.591041] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.598608] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251457, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.645032] env[61911]: DEBUG nova.compute.manager [req-9c0030f6-ceb6-4d33-99a6-b2ffe627d98f req-fcc081b8-4b3e-4de7-8a9d-34b13c47290a service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Received event network-vif-plugged-4863de42-7e22-4a3c-9024-c0d61f74e3f7 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 976.645114] env[61911]: DEBUG oslo_concurrency.lockutils [req-9c0030f6-ceb6-4d33-99a6-b2ffe627d98f req-fcc081b8-4b3e-4de7-8a9d-34b13c47290a service nova] Acquiring lock "3831af9a-fe2a-49d6-9e38-fc78e2616461-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.645303] env[61911]: DEBUG oslo_concurrency.lockutils [req-9c0030f6-ceb6-4d33-99a6-b2ffe627d98f req-fcc081b8-4b3e-4de7-8a9d-34b13c47290a service nova] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.645478] env[61911]: DEBUG oslo_concurrency.lockutils [req-9c0030f6-ceb6-4d33-99a6-b2ffe627d98f req-fcc081b8-4b3e-4de7-8a9d-34b13c47290a service nova] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.645648] env[61911]: DEBUG nova.compute.manager [req-9c0030f6-ceb6-4d33-99a6-b2ffe627d98f req-fcc081b8-4b3e-4de7-8a9d-34b13c47290a service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] No waiting events found dispatching network-vif-plugged-4863de42-7e22-4a3c-9024-c0d61f74e3f7 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 976.645820] env[61911]: WARNING nova.compute.manager [req-9c0030f6-ceb6-4d33-99a6-b2ffe627d98f req-fcc081b8-4b3e-4de7-8a9d-34b13c47290a service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Received unexpected event network-vif-plugged-4863de42-7e22-4a3c-9024-c0d61f74e3f7 for instance with vm_state building and task_state spawning. [ 976.726565] env[61911]: DEBUG nova.network.neutron [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Successfully updated port: 4863de42-7e22-4a3c-9024-c0d61f74e3f7 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.092801] env[61911]: DEBUG nova.scheduler.client.report [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.105920] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251457, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057104} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.106215] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.106995] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b269b45d-1187-41c3-99d9-2a7b482dc7a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.131155] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] d9234a5f-4d26-4bdc-8f32-3120830a8abe/d9234a5f-4d26-4bdc-8f32-3120830a8abe.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.131734] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9127b20-b189-4b7e-a055-7144be10dfde {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.151311] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 977.151311] env[61911]: value = "task-1251458" [ 977.151311] env[61911]: _type = "Task" [ 977.151311] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.159903] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251458, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.231904] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "refresh_cache-3831af9a-fe2a-49d6-9e38-fc78e2616461" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.232059] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "refresh_cache-3831af9a-fe2a-49d6-9e38-fc78e2616461" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.232214] env[61911]: DEBUG nova.network.neutron [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.366149] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "b594227d-1647-4e43-8c90-962a9e765388" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.366416] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "b594227d-1647-4e43-8c90-962a9e765388" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.603374] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.604021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.392s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.605481] env[61911]: INFO nova.compute.claims [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.611585] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "632e811e-b253-47ed-9d1e-6b2538ff804b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.611872] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.629545] env[61911]: INFO nova.scheduler.client.report [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Deleted allocations for instance 8051e8ec-6d82-416c-858a-f0cdd00f869d [ 977.662819] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251458, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.765764] env[61911]: DEBUG nova.network.neutron [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.873513] env[61911]: DEBUG nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 977.909855] env[61911]: DEBUG nova.network.neutron [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Updating instance_info_cache with network_info: [{"id": "4863de42-7e22-4a3c-9024-c0d61f74e3f7", "address": "fa:16:3e:27:1a:f1", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4863de42-7e", "ovs_interfaceid": "4863de42-7e22-4a3c-9024-c0d61f74e3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.114112] env[61911]: DEBUG nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 978.139016] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c7b447ac-9a50-461e-84a7-b443c69417cd tempest-ServersNegativeTestJSON-1948527724 tempest-ServersNegativeTestJSON-1948527724-project-member] Lock "8051e8ec-6d82-416c-858a-f0cdd00f869d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.439s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.163069] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251458, 'name': ReconfigVM_Task, 'duration_secs': 0.562215} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.163337] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Reconfigured VM instance instance-0000005b to attach disk [datastore1] d9234a5f-4d26-4bdc-8f32-3120830a8abe/d9234a5f-4d26-4bdc-8f32-3120830a8abe.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.164018] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4091babf-6523-42b3-8ec8-22fabc6cad77 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.170474] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 978.170474] env[61911]: value = "task-1251459" [ 978.170474] env[61911]: _type = "Task" [ 978.170474] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.178768] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251459, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.395447] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.412387] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "refresh_cache-3831af9a-fe2a-49d6-9e38-fc78e2616461" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.412716] env[61911]: DEBUG nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Instance network_info: |[{"id": "4863de42-7e22-4a3c-9024-c0d61f74e3f7", "address": "fa:16:3e:27:1a:f1", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4863de42-7e", "ovs_interfaceid": "4863de42-7e22-4a3c-9024-c0d61f74e3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 978.413552] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:1a:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4863de42-7e22-4a3c-9024-c0d61f74e3f7', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.421015] env[61911]: DEBUG oslo.service.loopingcall [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.421263] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.421498] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b65bce83-6603-4c15-a8c6-ec90eabe8902 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.441271] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.441271] env[61911]: value = "task-1251460" [ 978.441271] env[61911]: _type = "Task" [ 978.441271] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.448720] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251460, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.637164] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.669776] env[61911]: DEBUG nova.compute.manager [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Received event network-changed-4863de42-7e22-4a3c-9024-c0d61f74e3f7 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 978.669970] env[61911]: DEBUG nova.compute.manager [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Refreshing instance network info cache due to event network-changed-4863de42-7e22-4a3c-9024-c0d61f74e3f7. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 978.670215] env[61911]: DEBUG oslo_concurrency.lockutils [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] Acquiring lock "refresh_cache-3831af9a-fe2a-49d6-9e38-fc78e2616461" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.670360] env[61911]: DEBUG oslo_concurrency.lockutils [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] Acquired lock "refresh_cache-3831af9a-fe2a-49d6-9e38-fc78e2616461" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.670529] env[61911]: DEBUG nova.network.neutron [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Refreshing network info cache for port 4863de42-7e22-4a3c-9024-c0d61f74e3f7 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.684878] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251459, 'name': Rename_Task, 'duration_secs': 0.29918} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.685786] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.686183] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d41441c-2ada-4789-8107-f71033b7383b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.696424] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 978.696424] env[61911]: value = "task-1251461" [ 978.696424] env[61911]: _type = "Task" [ 978.696424] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.708165] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251461, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.821144] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1713ab-2991-4968-9239-41750d31eafb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.831793] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688fd14e-4f79-414c-a3ac-5ae8254e5f56 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.864089] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6b6af1-fb92-4493-aa13-d5a415585656 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.873167] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03815245-6f7c-4ea3-86e1-7eeb878bb0fc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.888997] env[61911]: DEBUG nova.compute.provider_tree [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.951369] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251460, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.209245] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251461, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.394417] env[61911]: DEBUG nova.scheduler.client.report [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.443198] env[61911]: DEBUG nova.network.neutron [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Updated VIF entry in instance network info cache for port 4863de42-7e22-4a3c-9024-c0d61f74e3f7. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 979.443198] env[61911]: DEBUG nova.network.neutron [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Updating instance_info_cache with network_info: [{"id": "4863de42-7e22-4a3c-9024-c0d61f74e3f7", "address": "fa:16:3e:27:1a:f1", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4863de42-7e", "ovs_interfaceid": "4863de42-7e22-4a3c-9024-c0d61f74e3f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.454917] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251460, 'name': CreateVM_Task, 'duration_secs': 0.567146} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.455086] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.455743] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.456544] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.456544] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.456544] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e509ed82-bfbd-4caf-80b7-35fd2ec09178 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.462080] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 979.462080] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5295e258-efb6-33b0-06d1-8b2c5171ab55" [ 979.462080] env[61911]: _type = "Task" [ 979.462080] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.471190] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5295e258-efb6-33b0-06d1-8b2c5171ab55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.709896] env[61911]: DEBUG oslo_vmware.api [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251461, 'name': PowerOnVM_Task, 'duration_secs': 0.527457} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.710316] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.710632] env[61911]: INFO nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Took 7.94 seconds to spawn the instance on the hypervisor. [ 979.710782] env[61911]: DEBUG nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 979.711657] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f5ff3f-ec2a-4a67-9c39-a791a7da9153 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.901298] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.901880] env[61911]: DEBUG nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 979.904574] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.860s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.949140] env[61911]: DEBUG oslo_concurrency.lockutils [req-2724482f-7466-4079-8ffd-ee3d418fb3a6 req-a150622e-79c9-4d5d-8b45-92b7d8511edb service nova] Releasing lock "refresh_cache-3831af9a-fe2a-49d6-9e38-fc78e2616461" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.974532] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5295e258-efb6-33b0-06d1-8b2c5171ab55, 'name': SearchDatastore_Task, 'duration_secs': 0.050321} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.974532] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.974757] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.974835] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.974973] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.975184] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.975776] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84dfaa8d-06e3-46e3-9f68-c5e094ef82bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.987480] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.987480] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.987910] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cf45a95-f3b1-4d57-aae9-a34b7d381f8c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.993850] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 979.993850] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ff81f9-a5d0-6133-424c-1767bc5e3e94" [ 979.993850] env[61911]: _type = "Task" [ 979.993850] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.002229] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ff81f9-a5d0-6133-424c-1767bc5e3e94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.233045] env[61911]: INFO nova.compute.manager [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Took 20.92 seconds to build instance. [ 980.408796] env[61911]: DEBUG nova.compute.utils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.412948] env[61911]: DEBUG nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 980.413105] env[61911]: DEBUG nova.network.neutron [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 980.471108] env[61911]: DEBUG nova.policy [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a5252f9309f4204a3a3fd7d865db6e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b285793da304b5e8f9736098ef163c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.514026] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ff81f9-a5d0-6133-424c-1767bc5e3e94, 'name': SearchDatastore_Task, 'duration_secs': 0.020621} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.514026] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-563dbd54-4777-4d44-b67a-12b8d2765202 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.520759] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 980.520759] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb91b7-426d-6ca6-1b86-fdf20efce90d" [ 980.520759] env[61911]: _type = "Task" [ 980.520759] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.531463] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb91b7-426d-6ca6-1b86-fdf20efce90d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.653169] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e93adc-9b11-4d9e-9200-c2be8eed871c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.661103] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942f77a2-cb59-435d-953b-38dfb0c6cd21 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.693129] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab7dd30-06f0-46b8-ab92-527a489f400a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.700506] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e4907c-2b5a-4aa4-a7b7-6d95d56e54aa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.713802] env[61911]: DEBUG nova.compute.provider_tree [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.734597] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6bdb5e58-89a6-437e-9255-52064309d158 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.431s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.818557] env[61911]: DEBUG nova.network.neutron [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Successfully created port: 7438c9cb-caf9-4f7d-b9b9-b9cba545a574 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.916094] env[61911]: DEBUG nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 981.032179] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fb91b7-426d-6ca6-1b86-fdf20efce90d, 'name': SearchDatastore_Task, 'duration_secs': 0.025989} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.032491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.032787] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 3831af9a-fe2a-49d6-9e38-fc78e2616461/3831af9a-fe2a-49d6-9e38-fc78e2616461.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 981.033082] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7231aaf-5d39-48b7-8f80-8711ef0a1ff7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.040534] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 981.040534] env[61911]: value = "task-1251462" [ 981.040534] env[61911]: _type = "Task" [ 981.040534] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.049167] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.217698] env[61911]: DEBUG nova.scheduler.client.report [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 981.551032] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251462, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.938661] env[61911]: DEBUG nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 981.965789] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 981.966042] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.966208] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 981.966395] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.966547] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 981.966697] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 981.966927] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 981.967108] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 981.967282] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 981.967449] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 981.967624] env[61911]: DEBUG nova.virt.hardware [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 981.968537] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-510bfc2c-8b77-4d23-b118-52d62d3e8bc1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.977062] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf69cd3-f507-44f2-bedf-de93ebc9bf70 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.051281] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515009} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.051484] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 3831af9a-fe2a-49d6-9e38-fc78e2616461/3831af9a-fe2a-49d6-9e38-fc78e2616461.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 982.051711] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.052043] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea2e1ae7-9e9f-4522-8d43-4c0e88e4d638 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.060846] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 982.060846] env[61911]: value = "task-1251463" [ 982.060846] env[61911]: _type = "Task" [ 982.060846] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.070491] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251463, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.229609] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.325s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.233530] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.838s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.236062] env[61911]: INFO nova.compute.claims [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.571076] env[61911]: DEBUG nova.compute.manager [req-93f5fee4-9439-4ec2-9a8e-bd6ea1101440 req-7e289b69-4d50-4b64-8c3f-7002275e09d5 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Received event network-vif-plugged-7438c9cb-caf9-4f7d-b9b9-b9cba545a574 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 982.571365] env[61911]: DEBUG oslo_concurrency.lockutils [req-93f5fee4-9439-4ec2-9a8e-bd6ea1101440 req-7e289b69-4d50-4b64-8c3f-7002275e09d5 service nova] Acquiring lock "f63fa31b-8840-4f95-9514-51f7d5393b25-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.571632] env[61911]: DEBUG oslo_concurrency.lockutils [req-93f5fee4-9439-4ec2-9a8e-bd6ea1101440 req-7e289b69-4d50-4b64-8c3f-7002275e09d5 service nova] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.571679] env[61911]: DEBUG oslo_concurrency.lockutils [req-93f5fee4-9439-4ec2-9a8e-bd6ea1101440 req-7e289b69-4d50-4b64-8c3f-7002275e09d5 service nova] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.571862] env[61911]: DEBUG nova.compute.manager [req-93f5fee4-9439-4ec2-9a8e-bd6ea1101440 req-7e289b69-4d50-4b64-8c3f-7002275e09d5 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] No waiting events found dispatching network-vif-plugged-7438c9cb-caf9-4f7d-b9b9-b9cba545a574 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 982.572105] env[61911]: WARNING nova.compute.manager [req-93f5fee4-9439-4ec2-9a8e-bd6ea1101440 req-7e289b69-4d50-4b64-8c3f-7002275e09d5 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Received unexpected event network-vif-plugged-7438c9cb-caf9-4f7d-b9b9-b9cba545a574 for instance with vm_state building and task_state spawning. [ 982.578720] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251463, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068018} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.578992] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.580032] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b75ca18-e3c5-4489-8875-ba920cc0699e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.608586] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 3831af9a-fe2a-49d6-9e38-fc78e2616461/3831af9a-fe2a-49d6-9e38-fc78e2616461.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.608940] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd1a6194-ebf9-4902-b741-5ce4ef0c1e17 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.631362] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 982.631362] env[61911]: value = "task-1251464" [ 982.631362] env[61911]: _type = "Task" [ 982.631362] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.640503] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251464, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.677483] env[61911]: DEBUG nova.network.neutron [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Successfully updated port: 7438c9cb-caf9-4f7d-b9b9-b9cba545a574 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.682492] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.682642] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.682823] env[61911]: DEBUG nova.compute.manager [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 982.683731] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fb57da-5d82-4bea-acf3-dd39d9d0c7a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.690769] env[61911]: DEBUG nova.compute.manager [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61911) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 982.691374] env[61911]: DEBUG nova.objects.instance [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lazy-loading 'flavor' on Instance uuid b1d2b205-81d7-410b-901d-ccd608bc5bec {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.796254] env[61911]: INFO nova.scheduler.client.report [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted allocation for migration 2e712027-ffdf-4233-8dba-1d056402fa73 [ 983.142870] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251464, 'name': ReconfigVM_Task, 'duration_secs': 0.375068} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.143640] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 3831af9a-fe2a-49d6-9e38-fc78e2616461/3831af9a-fe2a-49d6-9e38-fc78e2616461.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.144495] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d1c41a7-e563-499a-a6b1-2af2b3a21681 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.152058] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 983.152058] env[61911]: value = "task-1251465" [ 983.152058] env[61911]: _type = "Task" [ 983.152058] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.161258] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251465, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.180216] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-f63fa31b-8840-4f95-9514-51f7d5393b25" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.180481] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-f63fa31b-8840-4f95-9514-51f7d5393b25" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.180735] env[61911]: DEBUG nova.network.neutron [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 983.302749] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4f18640d-e584-4ade-8a79-dfcc05053557 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.113s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.422186] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c13845-dcc5-4409-8942-8cc82a36cd76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.430602] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3380a955-d667-4486-b7c1-074893b7fa3c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.461149] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7b50c3-6fe6-4100-969d-1db75ceab7f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.468731] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2f704c-7f07-49eb-8cdf-5a8b59e70476 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.482214] env[61911]: DEBUG nova.compute.provider_tree [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.661890] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251465, 'name': Rename_Task, 'duration_secs': 0.155738} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.662360] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.662498] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7f2ed6f-adb7-4897-8803-ba723dce7719 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.670250] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 983.670250] env[61911]: value = "task-1251466" [ 983.670250] env[61911]: _type = "Task" [ 983.670250] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.678374] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251466, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.698459] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.699218] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea2f1ad2-db93-4e9e-ac4e-067b007059b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.706230] env[61911]: DEBUG oslo_vmware.api [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 983.706230] env[61911]: value = "task-1251467" [ 983.706230] env[61911]: _type = "Task" [ 983.706230] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.717533] env[61911]: DEBUG oslo_vmware.api [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.718398] env[61911]: DEBUG nova.network.neutron [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 983.968413] env[61911]: DEBUG nova.network.neutron [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Updating instance_info_cache with network_info: [{"id": "7438c9cb-caf9-4f7d-b9b9-b9cba545a574", "address": "fa:16:3e:64:15:14", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7438c9cb-ca", "ovs_interfaceid": "7438c9cb-caf9-4f7d-b9b9-b9cba545a574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.985189] env[61911]: DEBUG nova.scheduler.client.report [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 984.180414] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251466, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.219390] env[61911]: DEBUG oslo_vmware.api [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251467, 'name': PowerOffVM_Task, 'duration_secs': 0.304036} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.219702] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.219967] env[61911]: DEBUG nova.compute.manager [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 984.220961] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4e17e7-7dde-4543-ba1b-76d085f06f92 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.446789] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f386424d-953d-431d-a56b-8542f2a3458c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.447080] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.447319] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f386424d-953d-431d-a56b-8542f2a3458c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.447490] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.447664] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.449735] env[61911]: INFO nova.compute.manager [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Terminating instance [ 984.471195] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-f63fa31b-8840-4f95-9514-51f7d5393b25" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.471514] env[61911]: DEBUG nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Instance network_info: |[{"id": "7438c9cb-caf9-4f7d-b9b9-b9cba545a574", "address": "fa:16:3e:64:15:14", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7438c9cb-ca", "ovs_interfaceid": "7438c9cb-caf9-4f7d-b9b9-b9cba545a574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 984.471864] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:15:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7438c9cb-caf9-4f7d-b9b9-b9cba545a574', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.483927] env[61911]: DEBUG oslo.service.loopingcall [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.484568] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.484922] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aab0bfd8-d14d-4140-8339-5e91c6986adf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.512044] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.512826] env[61911]: DEBUG nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 984.518494] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.882s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.519911] env[61911]: INFO nova.compute.claims [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.530496] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.530496] env[61911]: value = "task-1251468" [ 984.530496] env[61911]: _type = "Task" [ 984.530496] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.543996] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251468, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.598199] env[61911]: DEBUG nova.compute.manager [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Received event network-changed-7438c9cb-caf9-4f7d-b9b9-b9cba545a574 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 984.598423] env[61911]: DEBUG nova.compute.manager [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Refreshing instance network info cache due to event network-changed-7438c9cb-caf9-4f7d-b9b9-b9cba545a574. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 984.598629] env[61911]: DEBUG oslo_concurrency.lockutils [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] Acquiring lock "refresh_cache-f63fa31b-8840-4f95-9514-51f7d5393b25" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.598781] env[61911]: DEBUG oslo_concurrency.lockutils [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] Acquired lock "refresh_cache-f63fa31b-8840-4f95-9514-51f7d5393b25" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.598909] env[61911]: DEBUG nova.network.neutron [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Refreshing network info cache for port 7438c9cb-caf9-4f7d-b9b9-b9cba545a574 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 984.796815] env[61911]: DEBUG oslo_vmware.api [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251466, 'name': PowerOnVM_Task, 'duration_secs': 0.570741} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.796815] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.796815] env[61911]: INFO nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Took 8.73 seconds to spawn the instance on the hypervisor. [ 984.796815] env[61911]: DEBUG nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 984.796815] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07729019-8637-4091-8591-1d0e1f635011 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.796815] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9e7ffd20-fac6-4963-ad02-6a083aafa077 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.052s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.891993] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f44c597d-ffb8-446b-b814-65733d077dbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.892272] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f44c597d-ffb8-446b-b814-65733d077dbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.953806] env[61911]: DEBUG nova.compute.manager [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 984.954044] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.954950] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89649a41-f2d2-4758-a881-8cbca83a7a87 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.963722] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.964049] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-616a39ae-a0cc-4c20-8a5c-4bf55021d719 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.970689] env[61911]: DEBUG oslo_vmware.api [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 984.970689] env[61911]: value = "task-1251469" [ 984.970689] env[61911]: _type = "Task" [ 984.970689] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.978571] env[61911]: DEBUG oslo_vmware.api [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251469, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.023676] env[61911]: DEBUG nova.compute.utils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.028102] env[61911]: DEBUG nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 985.028102] env[61911]: DEBUG nova.network.neutron [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.040371] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251468, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.104867] env[61911]: DEBUG nova.policy [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cafb94e6eec44a109d46d806b6efbb90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6aa25bafac564a9db00225a42489babe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.203170] env[61911]: INFO nova.compute.manager [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Took 21.59 seconds to build instance. [ 985.358838] env[61911]: DEBUG nova.network.neutron [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Updated VIF entry in instance network info cache for port 7438c9cb-caf9-4f7d-b9b9-b9cba545a574. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 985.358838] env[61911]: DEBUG nova.network.neutron [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Updating instance_info_cache with network_info: [{"id": "7438c9cb-caf9-4f7d-b9b9-b9cba545a574", "address": "fa:16:3e:64:15:14", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7438c9cb-ca", "ovs_interfaceid": "7438c9cb-caf9-4f7d-b9b9-b9cba545a574", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.398442] env[61911]: DEBUG nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 985.483434] env[61911]: DEBUG oslo_vmware.api [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251469, 'name': PowerOffVM_Task, 'duration_secs': 0.211916} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.483734] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.483916] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 985.484194] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-feefda7d-1f96-4800-b649-f1e9f87f1291 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.530937] env[61911]: DEBUG nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 985.554213] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251468, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.558027] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.558027] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.558027] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleting the datastore file [datastore1] f386424d-953d-431d-a56b-8542f2a3458c {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.558027] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84cf7ba2-b4c6-4ff4-8313-08b57863e4d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.567478] env[61911]: DEBUG oslo_vmware.api [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 985.567478] env[61911]: value = "task-1251471" [ 985.567478] env[61911]: _type = "Task" [ 985.567478] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.575564] env[61911]: DEBUG nova.objects.instance [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lazy-loading 'flavor' on Instance uuid b1d2b205-81d7-410b-901d-ccd608bc5bec {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.580360] env[61911]: DEBUG oslo_vmware.api [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251471, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.661659] env[61911]: DEBUG nova.network.neutron [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Successfully created port: 681e329b-e012-42f2-bb1c-d1c63369e251 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.705522] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7ccde033-8831-4ed2-9457-31b296c6789b tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.102s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.769700] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38612e94-b0d5-4fe0-af5f-eac24ed148c0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.778803] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86de9069-d2ca-42b0-a108-f36b3a4e8a46 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.809142] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6c54e8-a389-49a4-9212-c890f7740006 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.818224] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69aa3c49-ae0f-4764-ae6c-c1441b1a1afe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.832863] env[61911]: DEBUG nova.compute.provider_tree [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.864991] env[61911]: DEBUG oslo_concurrency.lockutils [req-a3d7d1fc-c269-4f5f-a242-5ef8b23d5c86 req-9b10e065-be92-4a50-af66-426920ee1d79 service nova] Releasing lock "refresh_cache-f63fa31b-8840-4f95-9514-51f7d5393b25" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.915856] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.048024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "3831af9a-fe2a-49d6-9e38-fc78e2616461" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.048024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.048024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "3831af9a-fe2a-49d6-9e38-fc78e2616461-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.048024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.048024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.048024] env[61911]: INFO nova.compute.manager [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Terminating instance [ 986.053473] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251468, 'name': CreateVM_Task, 'duration_secs': 1.120043} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.053799] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.054570] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.054784] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.055181] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.055868] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db2a0d00-754a-4524-a85a-995632de3851 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.061944] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 986.061944] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52940f47-3f11-00a6-4479-d1643041c0cc" [ 986.061944] env[61911]: _type = "Task" [ 986.061944] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.072619] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52940f47-3f11-00a6-4479-d1643041c0cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.080418] env[61911]: DEBUG oslo_vmware.api [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251471, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.084015] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.084190] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquired lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.084378] env[61911]: DEBUG nova.network.neutron [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.084559] env[61911]: DEBUG nova.objects.instance [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lazy-loading 'info_cache' on Instance uuid b1d2b205-81d7-410b-901d-ccd608bc5bec {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.335637] env[61911]: DEBUG nova.scheduler.client.report [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 986.549029] env[61911]: DEBUG nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 986.555132] env[61911]: DEBUG nova.compute.manager [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 986.555356] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 986.556323] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf49b5a-779b-4ac4-ab4b-3bc7d7ed1cdf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.564557] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 986.567660] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf0a1e53-1e6e-4e9c-9ca1-2596bd6e6a42 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.578130] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 986.578441] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.578655] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 986.578897] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.579156] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 986.579409] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 986.579700] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 986.579942] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 986.580159] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 986.580338] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 986.580557] env[61911]: DEBUG nova.virt.hardware [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 986.586155] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab13b20-6587-4687-944d-f6ae7a8d3cdb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.588967] env[61911]: DEBUG oslo_vmware.api [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 986.588967] env[61911]: value = "task-1251472" [ 986.588967] env[61911]: _type = "Task" [ 986.588967] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.589267] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52940f47-3f11-00a6-4479-d1643041c0cc, 'name': SearchDatastore_Task, 'duration_secs': 0.01646} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.590971] env[61911]: DEBUG nova.objects.base [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 986.592349] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.592618] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.592919] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.593145] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.593370] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.600512] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40a720ee-8b16-4378-852f-0002603ccdc0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.602368] env[61911]: DEBUG oslo_vmware.api [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251471, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.606483} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.605212] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.605415] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.605597] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.605775] env[61911]: INFO nova.compute.manager [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Took 1.65 seconds to destroy the instance on the hypervisor. [ 986.606025] env[61911]: DEBUG oslo.service.loopingcall [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.607081] env[61911]: DEBUG nova.compute.manager [-] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 986.607206] env[61911]: DEBUG nova.network.neutron [-] [instance: f386424d-953d-431d-a56b-8542f2a3458c] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 986.609803] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05f6f6b-6d79-4a0a-a103-536f43a808b7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.617248] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.617426] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.618145] env[61911]: DEBUG oslo_vmware.api [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.618899] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-988023eb-6695-4618-9602-6e78e4cd9102 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.632351] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 986.632351] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5225c1f2-a772-2d84-6dbf-0470d55c8035" [ 986.632351] env[61911]: _type = "Task" [ 986.632351] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.640261] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5225c1f2-a772-2d84-6dbf-0470d55c8035, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.840923] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.841497] env[61911]: DEBUG nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 986.844557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.929s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.846105] env[61911]: INFO nova.compute.claims [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.898678] env[61911]: DEBUG nova.compute.manager [req-fa7ff273-ffa2-4e73-8a99-b1ffa1bb7aff req-26d59155-34ce-4fd9-86ad-927410398e60 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Received event network-vif-deleted-5937b7a1-0418-4e62-9605-d52f38ca02b4 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 986.898969] env[61911]: INFO nova.compute.manager [req-fa7ff273-ffa2-4e73-8a99-b1ffa1bb7aff req-26d59155-34ce-4fd9-86ad-927410398e60 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Neutron deleted interface 5937b7a1-0418-4e62-9605-d52f38ca02b4; detaching it from the instance and deleting it from the info cache [ 986.899077] env[61911]: DEBUG nova.network.neutron [req-fa7ff273-ffa2-4e73-8a99-b1ffa1bb7aff req-26d59155-34ce-4fd9-86ad-927410398e60 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.074024] env[61911]: DEBUG nova.compute.manager [req-3ec913f7-9dd6-4dfb-ad04-fc1631a82b1a req-59ab8f4b-0c65-4577-b2a6-b186e822fe7a service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Received event network-vif-plugged-681e329b-e012-42f2-bb1c-d1c63369e251 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 987.074024] env[61911]: DEBUG oslo_concurrency.lockutils [req-3ec913f7-9dd6-4dfb-ad04-fc1631a82b1a req-59ab8f4b-0c65-4577-b2a6-b186e822fe7a service nova] Acquiring lock "b594227d-1647-4e43-8c90-962a9e765388-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.074024] env[61911]: DEBUG oslo_concurrency.lockutils [req-3ec913f7-9dd6-4dfb-ad04-fc1631a82b1a req-59ab8f4b-0c65-4577-b2a6-b186e822fe7a service nova] Lock "b594227d-1647-4e43-8c90-962a9e765388-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.074024] env[61911]: DEBUG oslo_concurrency.lockutils [req-3ec913f7-9dd6-4dfb-ad04-fc1631a82b1a req-59ab8f4b-0c65-4577-b2a6-b186e822fe7a service nova] Lock "b594227d-1647-4e43-8c90-962a9e765388-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.074024] env[61911]: DEBUG nova.compute.manager [req-3ec913f7-9dd6-4dfb-ad04-fc1631a82b1a req-59ab8f4b-0c65-4577-b2a6-b186e822fe7a service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] No waiting events found dispatching network-vif-plugged-681e329b-e012-42f2-bb1c-d1c63369e251 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 987.074024] env[61911]: WARNING nova.compute.manager [req-3ec913f7-9dd6-4dfb-ad04-fc1631a82b1a req-59ab8f4b-0c65-4577-b2a6-b186e822fe7a service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Received unexpected event network-vif-plugged-681e329b-e012-42f2-bb1c-d1c63369e251 for instance with vm_state building and task_state spawning. [ 987.104847] env[61911]: DEBUG oslo_vmware.api [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251472, 'name': PowerOffVM_Task, 'duration_secs': 0.173618} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.104847] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 987.104847] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.104847] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f74a0aa5-4f75-45b6-b9c2-203e53dc81a0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.143487] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5225c1f2-a772-2d84-6dbf-0470d55c8035, 'name': SearchDatastore_Task, 'duration_secs': 0.013546} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.143487] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8260329-158f-4c2c-86ab-2ea6982ad94b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.148860] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 987.148860] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521d17b8-d9b3-ec9d-039b-c6ac649116b8" [ 987.148860] env[61911]: _type = "Task" [ 987.148860] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.157388] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521d17b8-d9b3-ec9d-039b-c6ac649116b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.171304] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.171304] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.171612] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleting the datastore file [datastore1] 3831af9a-fe2a-49d6-9e38-fc78e2616461 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.172573] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-286cc51d-4f21-438b-948f-e5a67d53d1f5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.180472] env[61911]: DEBUG nova.network.neutron [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Successfully updated port: 681e329b-e012-42f2-bb1c-d1c63369e251 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.187201] env[61911]: DEBUG oslo_vmware.api [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 987.187201] env[61911]: value = "task-1251474" [ 987.187201] env[61911]: _type = "Task" [ 987.187201] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.197556] env[61911]: DEBUG oslo_vmware.api [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.321037] env[61911]: DEBUG nova.network.neutron [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Updating instance_info_cache with network_info: [{"id": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "address": "fa:16:3e:f9:7d:ad", "network": {"id": "09acb8d3-5571-443b-a7e9-83a326b9b90e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1824349544-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fe1ad57908647b79849aea9d81726a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450ba0e3-64", "ovs_interfaceid": "450ba0e3-64ce-4aff-bf9d-059d8aa237fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.349810] env[61911]: DEBUG nova.compute.utils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.353050] env[61911]: DEBUG nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 987.353714] env[61911]: DEBUG nova.network.neutron [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.374670] env[61911]: DEBUG nova.network.neutron [-] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.402473] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7eab738b-5d3c-451f-8d8f-942f8e0f16c0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.413180] env[61911]: DEBUG nova.policy [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cafb94e6eec44a109d46d806b6efbb90', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6aa25bafac564a9db00225a42489babe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 987.417556] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bfa512-13db-49d8-ae0b-de69fcc299ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.452134] env[61911]: DEBUG nova.compute.manager [req-fa7ff273-ffa2-4e73-8a99-b1ffa1bb7aff req-26d59155-34ce-4fd9-86ad-927410398e60 service nova] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Detach interface failed, port_id=5937b7a1-0418-4e62-9605-d52f38ca02b4, reason: Instance f386424d-953d-431d-a56b-8542f2a3458c could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 987.663363] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521d17b8-d9b3-ec9d-039b-c6ac649116b8, 'name': SearchDatastore_Task, 'duration_secs': 0.012308} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.663363] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.663363] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] f63fa31b-8840-4f95-9514-51f7d5393b25/f63fa31b-8840-4f95-9514-51f7d5393b25.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.663363] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-203ed7e6-cdfc-47c9-9e60-49f51b3ff8ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.671335] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 987.671335] env[61911]: value = "task-1251475" [ 987.671335] env[61911]: _type = "Task" [ 987.671335] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.681091] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251475, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.687687] env[61911]: DEBUG nova.network.neutron [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Successfully created port: 07ff77e7-946d-4de9-8e85-badc55c396c1 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.689806] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "refresh_cache-b594227d-1647-4e43-8c90-962a9e765388" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.689987] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired lock "refresh_cache-b594227d-1647-4e43-8c90-962a9e765388" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.690089] env[61911]: DEBUG nova.network.neutron [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.701250] env[61911]: DEBUG oslo_vmware.api [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.378344} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.702396] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.702556] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 987.702741] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 987.702926] env[61911]: INFO nova.compute.manager [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Took 1.15 seconds to destroy the instance on the hypervisor. [ 987.703232] env[61911]: DEBUG oslo.service.loopingcall [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.703684] env[61911]: DEBUG nova.compute.manager [-] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 987.703784] env[61911]: DEBUG nova.network.neutron [-] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 987.825238] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Releasing lock "refresh_cache-b1d2b205-81d7-410b-901d-ccd608bc5bec" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.858932] env[61911]: DEBUG nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 987.879633] env[61911]: INFO nova.compute.manager [-] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Took 1.27 seconds to deallocate network for instance. [ 987.991699] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 987.992271] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.117471] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2262047-5124-4039-82b2-ca97cd0bd7a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.127203] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09538b1e-1fd1-4b34-85a1-78487772de29 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.164510] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac35e042-9ddc-42d7-8a46-488ca300b458 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.172852] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b18b632-56d3-4283-b729-379fa3b5c1eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.885232] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.885631] env[61911]: DEBUG nova.network.neutron [-] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.891810] env[61911]: DEBUG nova.compute.provider_tree [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.897338] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.897338] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 988.897919] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251475, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494237} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.899738] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] f63fa31b-8840-4f95-9514-51f7d5393b25/f63fa31b-8840-4f95-9514-51f7d5393b25.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.899965] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.900354] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-351e9e99-b9cb-4569-906c-677ea00e7fc2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.910229] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 988.910229] env[61911]: value = "task-1251476" [ 988.910229] env[61911]: _type = "Task" [ 988.910229] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.921254] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251476, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.930885] env[61911]: DEBUG nova.network.neutron [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.095732] env[61911]: DEBUG nova.network.neutron [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Updating instance_info_cache with network_info: [{"id": "681e329b-e012-42f2-bb1c-d1c63369e251", "address": "fa:16:3e:21:bc:69", "network": {"id": "d8b29fff-4bb9-4212-8ffe-a9af2945bcad", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "70794eff7a2c478b800918ba4144a52b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap681e329b-e0", "ovs_interfaceid": "681e329b-e012-42f2-bb1c-d1c63369e251", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.114356] env[61911]: DEBUG nova.compute.manager [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Received event network-changed-681e329b-e012-42f2-bb1c-d1c63369e251 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 989.114356] env[61911]: DEBUG nova.compute.manager [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Refreshing instance network info cache due to event network-changed-681e329b-e012-42f2-bb1c-d1c63369e251. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 989.114356] env[61911]: DEBUG oslo_concurrency.lockutils [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] Acquiring lock "refresh_cache-b594227d-1647-4e43-8c90-962a9e765388" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.122841] env[61911]: DEBUG nova.compute.manager [req-1e259e55-d050-4cea-8906-ed068df85cfd req-de0080ae-3c06-4857-9079-5bbf4a986a96 service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Received event network-vif-plugged-07ff77e7-946d-4de9-8e85-badc55c396c1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 989.122930] env[61911]: DEBUG oslo_concurrency.lockutils [req-1e259e55-d050-4cea-8906-ed068df85cfd req-de0080ae-3c06-4857-9079-5bbf4a986a96 service nova] Acquiring lock "632e811e-b253-47ed-9d1e-6b2538ff804b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.123177] env[61911]: DEBUG oslo_concurrency.lockutils [req-1e259e55-d050-4cea-8906-ed068df85cfd req-de0080ae-3c06-4857-9079-5bbf4a986a96 service nova] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.123507] env[61911]: DEBUG oslo_concurrency.lockutils [req-1e259e55-d050-4cea-8906-ed068df85cfd req-de0080ae-3c06-4857-9079-5bbf4a986a96 service nova] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.123631] env[61911]: DEBUG nova.compute.manager [req-1e259e55-d050-4cea-8906-ed068df85cfd req-de0080ae-3c06-4857-9079-5bbf4a986a96 service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] No waiting events found dispatching network-vif-plugged-07ff77e7-946d-4de9-8e85-badc55c396c1 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 989.123755] env[61911]: WARNING nova.compute.manager [req-1e259e55-d050-4cea-8906-ed068df85cfd req-de0080ae-3c06-4857-9079-5bbf4a986a96 service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Received unexpected event network-vif-plugged-07ff77e7-946d-4de9-8e85-badc55c396c1 for instance with vm_state building and task_state spawning. [ 989.195822] env[61911]: DEBUG nova.network.neutron [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Successfully updated port: 07ff77e7-946d-4de9-8e85-badc55c396c1 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.387384] env[61911]: INFO nova.compute.manager [-] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Took 1.68 seconds to deallocate network for instance. [ 989.393484] env[61911]: DEBUG nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 989.395595] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.396279] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b747f9d-ad20-4c0a-b025-d467045e30e1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.399951] env[61911]: DEBUG nova.scheduler.client.report [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 989.408796] env[61911]: DEBUG oslo_vmware.api [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 989.408796] env[61911]: value = "task-1251477" [ 989.408796] env[61911]: _type = "Task" [ 989.408796] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.427382] env[61911]: DEBUG oslo_vmware.api [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.430638] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 989.430982] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.431242] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 989.431529] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.431762] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 989.432010] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 989.432334] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 989.432592] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 989.432859] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 989.433173] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 989.433402] env[61911]: DEBUG nova.virt.hardware [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 989.434931] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d51e67-e83a-44b3-9295-2ce56c091b39 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.439388] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.439558] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.439756] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 989.445600] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251476, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076492} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.446641] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.447776] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b25358-917d-4f4c-a37a-e63d93204500 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.457445] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659d20be-e04d-4d41-9e78-942e29f9000f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.481942] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] f63fa31b-8840-4f95-9514-51f7d5393b25/f63fa31b-8840-4f95-9514-51f7d5393b25.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.482949] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d9d9185-cff2-4b9d-9681-4f1a102eab05 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.511843] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 989.511843] env[61911]: value = "task-1251478" [ 989.511843] env[61911]: _type = "Task" [ 989.511843] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.519948] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251478, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.597290] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Releasing lock "refresh_cache-b594227d-1647-4e43-8c90-962a9e765388" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.597703] env[61911]: DEBUG nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Instance network_info: |[{"id": "681e329b-e012-42f2-bb1c-d1c63369e251", "address": "fa:16:3e:21:bc:69", "network": {"id": "d8b29fff-4bb9-4212-8ffe-a9af2945bcad", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "70794eff7a2c478b800918ba4144a52b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap681e329b-e0", "ovs_interfaceid": "681e329b-e012-42f2-bb1c-d1c63369e251", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 989.598061] env[61911]: DEBUG oslo_concurrency.lockutils [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] Acquired lock "refresh_cache-b594227d-1647-4e43-8c90-962a9e765388" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.598256] env[61911]: DEBUG nova.network.neutron [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Refreshing network info cache for port 681e329b-e012-42f2-bb1c-d1c63369e251 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.599479] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:bc:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cf63c3c8-d774-4b81-9b12-848612a96076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '681e329b-e012-42f2-bb1c-d1c63369e251', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.607467] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Creating folder: Project (6aa25bafac564a9db00225a42489babe). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 989.608036] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d83b37c6-d883-44f0-a667-dcae551c0e9e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.622298] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Created folder: Project (6aa25bafac564a9db00225a42489babe) in parent group-v269521. [ 989.622614] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Creating folder: Instances. Parent ref: group-v269653. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 989.622887] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-92c6c909-df1c-4299-b0ca-4a9b17fe45ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.634126] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Created folder: Instances in parent group-v269653. [ 989.634385] env[61911]: DEBUG oslo.service.loopingcall [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.634595] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b594227d-1647-4e43-8c90-962a9e765388] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.634814] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-86166a66-f636-4b26-8efe-6d771099bc56 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.655546] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.655546] env[61911]: value = "task-1251481" [ 989.655546] env[61911]: _type = "Task" [ 989.655546] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.664421] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251481, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.697924] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "refresh_cache-632e811e-b253-47ed-9d1e-6b2538ff804b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.698199] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired lock "refresh_cache-632e811e-b253-47ed-9d1e-6b2538ff804b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.698294] env[61911]: DEBUG nova.network.neutron [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.898144] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.910385] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.066s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.911050] env[61911]: DEBUG nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 989.914070] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.029s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.914205] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.916817] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.019s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.917044] env[61911]: DEBUG nova.objects.instance [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lazy-loading 'resources' on Instance uuid 3831af9a-fe2a-49d6-9e38-fc78e2616461 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.928014] env[61911]: DEBUG oslo_vmware.api [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251477, 'name': PowerOnVM_Task, 'duration_secs': 0.426659} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.928334] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.928562] env[61911]: DEBUG nova.compute.manager [None req-3fdd2a24-3c62-4e8d-a65d-8d090cca61fe tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 989.929476] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b06b2a9-e447-4c7d-8faa-3c5f444c8d0e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.942435] env[61911]: INFO nova.scheduler.client.report [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted allocations for instance f386424d-953d-431d-a56b-8542f2a3458c [ 990.021801] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251478, 'name': ReconfigVM_Task, 'duration_secs': 0.299523} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.022086] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Reconfigured VM instance instance-0000005d to attach disk [datastore1] f63fa31b-8840-4f95-9514-51f7d5393b25/f63fa31b-8840-4f95-9514-51f7d5393b25.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 990.022771] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f35fa35-86a9-47c2-af93-b39513c93a30 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.031575] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 990.031575] env[61911]: value = "task-1251482" [ 990.031575] env[61911]: _type = "Task" [ 990.031575] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.040491] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251482, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.165571] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251481, 'name': CreateVM_Task, 'duration_secs': 0.491698} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.165768] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b594227d-1647-4e43-8c90-962a9e765388] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.166430] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.166600] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.166926] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.167193] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc5184c4-ceab-449e-9631-f8f9ad30d93b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.171920] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 990.171920] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5296e2ff-225a-441d-f82e-1dd5e8a1008d" [ 990.171920] env[61911]: _type = "Task" [ 990.171920] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.180320] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5296e2ff-225a-441d-f82e-1dd5e8a1008d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.228536] env[61911]: DEBUG nova.network.neutron [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 990.306676] env[61911]: DEBUG nova.network.neutron [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Updated VIF entry in instance network info cache for port 681e329b-e012-42f2-bb1c-d1c63369e251. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.307070] env[61911]: DEBUG nova.network.neutron [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Updating instance_info_cache with network_info: [{"id": "681e329b-e012-42f2-bb1c-d1c63369e251", "address": "fa:16:3e:21:bc:69", "network": {"id": "d8b29fff-4bb9-4212-8ffe-a9af2945bcad", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "70794eff7a2c478b800918ba4144a52b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap681e329b-e0", "ovs_interfaceid": "681e329b-e012-42f2-bb1c-d1c63369e251", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.356696] env[61911]: DEBUG nova.network.neutron [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Updating instance_info_cache with network_info: [{"id": "07ff77e7-946d-4de9-8e85-badc55c396c1", "address": "fa:16:3e:47:3c:56", "network": {"id": "d8b29fff-4bb9-4212-8ffe-a9af2945bcad", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.225", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "70794eff7a2c478b800918ba4144a52b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07ff77e7-94", "ovs_interfaceid": "07ff77e7-946d-4de9-8e85-badc55c396c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.422612] env[61911]: DEBUG nova.compute.utils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 990.427407] env[61911]: DEBUG nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 990.427407] env[61911]: DEBUG nova.network.neutron [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 990.448520] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4acd8b28-45ff-4a03-b5f1-3fd10da04d71 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f386424d-953d-431d-a56b-8542f2a3458c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.466465] env[61911]: DEBUG nova.policy [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b034026eede4f5ca4757b552f7dca51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5be8ec1d59a14d618906ec32d0e134e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 990.547386] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251482, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.623944] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbac9dc-408e-43e7-bf2c-e63867d1fa61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.632292] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4953060f-bb93-46c8-b25c-f128d0e5485c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.664855] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2172075e-0951-406a-99ef-150a89056388 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.673199] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Updating instance_info_cache with network_info: [{"id": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "address": "fa:16:3e:59:fa:2e", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d69ab4c-3b", "ovs_interfaceid": "8d69ab4c-3b99-412f-8985-3d8a9210f776", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.678225] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6add7c79-d60a-4c24-882f-851b90250019 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.692425] env[61911]: DEBUG nova.compute.provider_tree [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.697288] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5296e2ff-225a-441d-f82e-1dd5e8a1008d, 'name': SearchDatastore_Task, 'duration_secs': 0.011864} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.697558] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.700296] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.700296] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.700296] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.700296] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.700296] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5bbf25ce-7167-428c-ae76-9fa50429248d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.708670] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.708853] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.709787] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ec72c8f-b898-4bde-81ee-64be0f9b8fc4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.717073] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 990.717073] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52793f00-0714-22c2-85d3-a24fd7fcea6c" [ 990.717073] env[61911]: _type = "Task" [ 990.717073] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.726032] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52793f00-0714-22c2-85d3-a24fd7fcea6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.809946] env[61911]: DEBUG oslo_concurrency.lockutils [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] Releasing lock "refresh_cache-b594227d-1647-4e43-8c90-962a9e765388" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.810292] env[61911]: DEBUG nova.compute.manager [req-b5ed04be-adb8-4178-86eb-4c2a92e6fabb req-481a8197-6b4e-416c-8ddb-83317c5b72e4 service nova] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Received event network-vif-deleted-4863de42-7e22-4a3c-9024-c0d61f74e3f7 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 990.859720] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Releasing lock "refresh_cache-632e811e-b253-47ed-9d1e-6b2538ff804b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.862531] env[61911]: DEBUG nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Instance network_info: |[{"id": "07ff77e7-946d-4de9-8e85-badc55c396c1", "address": "fa:16:3e:47:3c:56", "network": {"id": "d8b29fff-4bb9-4212-8ffe-a9af2945bcad", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.225", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "70794eff7a2c478b800918ba4144a52b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07ff77e7-94", "ovs_interfaceid": "07ff77e7-946d-4de9-8e85-badc55c396c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 990.862531] env[61911]: DEBUG nova.network.neutron [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Successfully created port: c2701d58-2bb9-447b-8b02-6f3c3e4cb5de {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.863084] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:3c:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cf63c3c8-d774-4b81-9b12-848612a96076', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07ff77e7-946d-4de9-8e85-badc55c396c1', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.870309] env[61911]: DEBUG oslo.service.loopingcall [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.870536] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 990.871112] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98b4c7de-e651-41b6-bbba-6b64cfeb50d8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.891584] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.891584] env[61911]: value = "task-1251483" [ 990.891584] env[61911]: _type = "Task" [ 990.891584] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.905170] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251483, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.930019] env[61911]: DEBUG nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 991.045301] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251482, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.178374] env[61911]: DEBUG nova.compute.manager [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Received event network-changed-07ff77e7-946d-4de9-8e85-badc55c396c1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 991.178577] env[61911]: DEBUG nova.compute.manager [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Refreshing instance network info cache due to event network-changed-07ff77e7-946d-4de9-8e85-badc55c396c1. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 991.178800] env[61911]: DEBUG oslo_concurrency.lockutils [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] Acquiring lock "refresh_cache-632e811e-b253-47ed-9d1e-6b2538ff804b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.178946] env[61911]: DEBUG oslo_concurrency.lockutils [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] Acquired lock "refresh_cache-632e811e-b253-47ed-9d1e-6b2538ff804b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.181048] env[61911]: DEBUG nova.network.neutron [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Refreshing network info cache for port 07ff77e7-946d-4de9-8e85-badc55c396c1 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.186309] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-31f0a4f7-a69d-4ec4-966d-93f402163b88" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.186309] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 991.186309] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.186309] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.186309] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.186309] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.186309] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.186309] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.186309] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 991.186309] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.198606] env[61911]: DEBUG nova.scheduler.client.report [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 991.228080] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52793f00-0714-22c2-85d3-a24fd7fcea6c, 'name': SearchDatastore_Task, 'duration_secs': 0.009685} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.228872] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08449a3e-2b02-4138-8d59-cc25a7790ad8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.235139] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 991.235139] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fba097-934d-e7fa-7647-9580ff4a26f3" [ 991.235139] env[61911]: _type = "Task" [ 991.235139] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.243147] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fba097-934d-e7fa-7647-9580ff4a26f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.404391] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251483, 'name': CreateVM_Task, 'duration_secs': 0.398497} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.404655] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 991.405498] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.405768] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.406203] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.406500] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e65a5cd-1fda-4610-9364-3257c0fbd033 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.412234] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 991.412234] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f0b666-cc90-6ff7-d6ef-07b107031626" [ 991.412234] env[61911]: _type = "Task" [ 991.412234] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.421523] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f0b666-cc90-6ff7-d6ef-07b107031626, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.545097] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251482, 'name': Rename_Task, 'duration_secs': 1.177021} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.545367] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 991.545620] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f61f61ab-ed69-4319-8c2e-c62659e7e672 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.552813] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 991.552813] env[61911]: value = "task-1251484" [ 991.552813] env[61911]: _type = "Task" [ 991.552813] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.560569] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251484, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.687193] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.704650] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.788s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.707259] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.020s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.707637] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.707637] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 991.709171] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b49266-9fc4-4c31-9905-ff2123567fdf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.717906] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b52c96-3e9f-4a90-b8f1-419131d12d75 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.739373] env[61911]: INFO nova.scheduler.client.report [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted allocations for instance 3831af9a-fe2a-49d6-9e38-fc78e2616461 [ 991.741211] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95235c05-fbd2-436b-85e9-4c8f30fcea7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.756025] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52fba097-934d-e7fa-7647-9580ff4a26f3, 'name': SearchDatastore_Task, 'duration_secs': 0.010569} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.758251] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.759220] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] b594227d-1647-4e43-8c90-962a9e765388/b594227d-1647-4e43-8c90-962a9e765388.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.759220] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8f298d8-387e-4b65-bd77-79f91a019e81 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.761822] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7e11e4-8150-42c5-a048-fb862e939c76 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.797476] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 991.797476] env[61911]: value = "task-1251485" [ 991.797476] env[61911]: _type = "Task" [ 991.797476] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.797868] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180526MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 991.797954] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.798133] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.811846] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251485, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.927022] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f0b666-cc90-6ff7-d6ef-07b107031626, 'name': SearchDatastore_Task, 'duration_secs': 0.010121} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.929694] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.930491] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.930944] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.931315] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.931723] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.932215] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5cb09e80-de39-4a63-80a7-349a539ec96a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.943024] env[61911]: DEBUG nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 991.944057] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.944435] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 991.945793] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83dcf07c-af12-4824-80b3-7f34e1fc7ce2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.953062] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 991.953062] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221fbc8-e10a-1bbe-fadd-e7d2c7216716" [ 991.953062] env[61911]: _type = "Task" [ 991.953062] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.966377] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221fbc8-e10a-1bbe-fadd-e7d2c7216716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.975067] env[61911]: DEBUG nova.network.neutron [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Updated VIF entry in instance network info cache for port 07ff77e7-946d-4de9-8e85-badc55c396c1. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.975067] env[61911]: DEBUG nova.network.neutron [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Updating instance_info_cache with network_info: [{"id": "07ff77e7-946d-4de9-8e85-badc55c396c1", "address": "fa:16:3e:47:3c:56", "network": {"id": "d8b29fff-4bb9-4212-8ffe-a9af2945bcad", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.225", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "70794eff7a2c478b800918ba4144a52b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf63c3c8-d774-4b81-9b12-848612a96076", "external-id": "nsx-vlan-transportzone-315", "segmentation_id": 315, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07ff77e7-94", "ovs_interfaceid": "07ff77e7-946d-4de9-8e85-badc55c396c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.980586] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 991.980586] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.980586] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 991.980586] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.981127] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 991.981497] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 991.981870] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 991.982085] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 991.982309] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 991.982548] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 991.982836] env[61911]: DEBUG nova.virt.hardware [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 991.984741] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aea15b6-032b-4eb3-868b-c5128f2b7d1d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.991080] env[61911]: DEBUG oslo_concurrency.lockutils [req-48debbb4-8964-4bee-a55d-96a63da0df5b req-70a06ace-847b-468a-b5c4-d6510fe79e7e service nova] Releasing lock "refresh_cache-632e811e-b253-47ed-9d1e-6b2538ff804b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.002421] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea40c68b-0db6-44a3-a3d0-251d88ebf4bc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.065727] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251484, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.253810] env[61911]: DEBUG oslo_concurrency.lockutils [None req-23de172f-e6a4-4124-874c-8d8704bae615 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "3831af9a-fe2a-49d6-9e38-fc78e2616461" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.209s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.318459] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251485, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.464879] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5221fbc8-e10a-1bbe-fadd-e7d2c7216716, 'name': SearchDatastore_Task, 'duration_secs': 0.010317} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.465780] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54b0b308-4139-4a31-8405-e7f091528173 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.471794] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 992.471794] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5207a1ae-21e3-c2df-6df2-aebeaff0a8f9" [ 992.471794] env[61911]: _type = "Task" [ 992.471794] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.480648] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5207a1ae-21e3-c2df-6df2-aebeaff0a8f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.501770] env[61911]: DEBUG nova.compute.manager [req-ba678469-02e5-4253-bd74-a63e74c6affd req-0f121d7f-ec39-4c22-87c0-c3956440629c service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Received event network-vif-plugged-c2701d58-2bb9-447b-8b02-6f3c3e4cb5de {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 992.502094] env[61911]: DEBUG oslo_concurrency.lockutils [req-ba678469-02e5-4253-bd74-a63e74c6affd req-0f121d7f-ec39-4c22-87c0-c3956440629c service nova] Acquiring lock "f44c597d-ffb8-446b-b814-65733d077dbf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.502238] env[61911]: DEBUG oslo_concurrency.lockutils [req-ba678469-02e5-4253-bd74-a63e74c6affd req-0f121d7f-ec39-4c22-87c0-c3956440629c service nova] Lock "f44c597d-ffb8-446b-b814-65733d077dbf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.502481] env[61911]: DEBUG oslo_concurrency.lockutils [req-ba678469-02e5-4253-bd74-a63e74c6affd req-0f121d7f-ec39-4c22-87c0-c3956440629c service nova] Lock "f44c597d-ffb8-446b-b814-65733d077dbf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.502662] env[61911]: DEBUG nova.compute.manager [req-ba678469-02e5-4253-bd74-a63e74c6affd req-0f121d7f-ec39-4c22-87c0-c3956440629c service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] No waiting events found dispatching network-vif-plugged-c2701d58-2bb9-447b-8b02-6f3c3e4cb5de {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 992.502835] env[61911]: WARNING nova.compute.manager [req-ba678469-02e5-4253-bd74-a63e74c6affd req-0f121d7f-ec39-4c22-87c0-c3956440629c service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Received unexpected event network-vif-plugged-c2701d58-2bb9-447b-8b02-6f3c3e4cb5de for instance with vm_state building and task_state spawning. [ 992.564161] env[61911]: DEBUG oslo_vmware.api [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251484, 'name': PowerOnVM_Task, 'duration_secs': 0.522188} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.564474] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.564704] env[61911]: INFO nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Took 10.63 seconds to spawn the instance on the hypervisor. [ 992.564892] env[61911]: DEBUG nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 992.565668] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95d9905-f7b8-4a35-8f8f-8c18f9456916 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.627963] env[61911]: DEBUG oslo_concurrency.lockutils [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.628256] env[61911]: DEBUG oslo_concurrency.lockutils [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.658911] env[61911]: DEBUG nova.network.neutron [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Successfully updated port: c2701d58-2bb9-447b-8b02-6f3c3e4cb5de {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.817415] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251485, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.830671] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 31f0a4f7-a69d-4ec4-966d-93f402163b88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.830738] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bafd58bc-8ae7-49a4-a039-31d328699010 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.830852] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance c95707ca-7e32-42c0-914a-f0b178f07997 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.830974] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5b7d09e1-ee14-4001-8d73-14e763402670 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831131] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831254] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance fddba36d-1b15-43fb-9e99-68880e8a235c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831368] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b1d2b205-81d7-410b-901d-ccd608bc5bec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831480] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831589] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance d9234a5f-4d26-4bdc-8f32-3120830a8abe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831698] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance f63fa31b-8840-4f95-9514-51f7d5393b25 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831805] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance b594227d-1647-4e43-8c90-962a9e765388 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.831916] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 632e811e-b253-47ed-9d1e-6b2538ff804b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.832038] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance f44c597d-ffb8-446b-b814-65733d077dbf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.982837] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5207a1ae-21e3-c2df-6df2-aebeaff0a8f9, 'name': SearchDatastore_Task, 'duration_secs': 0.035036} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.983240] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.983240] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 632e811e-b253-47ed-9d1e-6b2538ff804b/632e811e-b253-47ed-9d1e-6b2538ff804b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.983478] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78def4d0-dfbc-4a84-847b-9039790b4b83 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.990705] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 992.990705] env[61911]: value = "task-1251486" [ 992.990705] env[61911]: _type = "Task" [ 992.990705] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.084936] env[61911]: INFO nova.compute.manager [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Took 18.89 seconds to build instance. [ 993.131242] env[61911]: INFO nova.compute.manager [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Detaching volume 8ade23e8-4f87-4c13-9e03-db70f4c4988d [ 993.150500] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.150774] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.151042] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.151254] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.151435] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.153728] env[61911]: INFO nova.compute.manager [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Terminating instance [ 993.163551] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "refresh_cache-f44c597d-ffb8-446b-b814-65733d077dbf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.163551] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "refresh_cache-f44c597d-ffb8-446b-b814-65733d077dbf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.163551] env[61911]: DEBUG nova.network.neutron [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.168589] env[61911]: INFO nova.virt.block_device [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Attempting to driver detach volume 8ade23e8-4f87-4c13-9e03-db70f4c4988d from mountpoint /dev/sdb [ 993.168836] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 993.169038] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269631', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'name': 'volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c', 'attached_at': '', 'detached_at': '', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'serial': '8ade23e8-4f87-4c13-9e03-db70f4c4988d'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 993.170720] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e00df2c-6e6e-48e6-a9fa-054bb1b57985 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.194206] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c4de01-eb2e-4bbc-9ccc-ffe6fec3e645 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.202528] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01b8b64-3d66-4d77-a33c-258f97b61dc9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.225639] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5c99d7-3b7d-48f3-a099-5bd380a807df {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.243168] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] The volume has not been displaced from its original location: [datastore1] volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d/volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 993.249334] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 993.249486] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdcff556-6094-4d63-90ec-f6f6d000f0dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.269331] env[61911]: DEBUG oslo_vmware.api [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 993.269331] env[61911]: value = "task-1251487" [ 993.269331] env[61911]: _type = "Task" [ 993.269331] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.277857] env[61911]: DEBUG oslo_vmware.api [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251487, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.287441] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "a03413fe-e74d-42e4-83ea-a9c19318526e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.287742] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.312838] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251485, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.334467] env[61911]: INFO nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance a03413fe-e74d-42e4-83ea-a9c19318526e has allocations against this compute host but is not found in the database. [ 993.334795] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 993.335011] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 993.503076] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.517479] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7dfe665-ac6e-463b-8e7a-3b1b91ef3175 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.525284] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b1c48-163e-4d22-8b2e-2c7ce300755c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.556363] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a97c88-13ff-41ca-9aba-52310a87d96a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.564989] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26708f35-3619-4a7d-8233-fa5c47ee7fd0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.580387] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.586638] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d5c71768-1e53-4426-823c-0fbf552925cd tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.400s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.658054] env[61911]: DEBUG nova.compute.manager [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 993.658306] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 993.659200] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13612598-4bb2-4f7d-87df-6c4a13e48c69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.668996] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.669558] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2677d766-4e49-4ad8-88f2-ff201aa7e2f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.676751] env[61911]: DEBUG oslo_vmware.api [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 993.676751] env[61911]: value = "task-1251488" [ 993.676751] env[61911]: _type = "Task" [ 993.676751] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.685269] env[61911]: DEBUG oslo_vmware.api [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251488, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.705555] env[61911]: DEBUG nova.network.neutron [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.780417] env[61911]: DEBUG oslo_vmware.api [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251487, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.790151] env[61911]: DEBUG nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 993.817195] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251485, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.616759} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.817542] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] b594227d-1647-4e43-8c90-962a9e765388/b594227d-1647-4e43-8c90-962a9e765388.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.817735] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.818012] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-415bf763-db4e-4b3e-8ee6-46e688de747f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.830044] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 993.830044] env[61911]: value = "task-1251489" [ 993.830044] env[61911]: _type = "Task" [ 993.830044] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.843183] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251489, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.963240] env[61911]: DEBUG nova.network.neutron [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Updating instance_info_cache with network_info: [{"id": "c2701d58-2bb9-447b-8b02-6f3c3e4cb5de", "address": "fa:16:3e:92:92:c0", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2701d58-2b", "ovs_interfaceid": "c2701d58-2bb9-447b-8b02-6f3c3e4cb5de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.001266] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251486, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.083570] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 994.186718] env[61911]: DEBUG oslo_vmware.api [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251488, 'name': PowerOffVM_Task, 'duration_secs': 0.278605} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.186997] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 994.187184] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.187439] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bced9569-9829-43fb-b147-61616164874c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.275642] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.275875] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.276068] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleting the datastore file [datastore1] d9234a5f-4d26-4bdc-8f32-3120830a8abe {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.276360] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e7cbaf6-3c72-42bf-840b-cc34fb7196ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.281317] env[61911]: DEBUG oslo_vmware.api [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251487, 'name': ReconfigVM_Task, 'duration_secs': 0.577707} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.281960] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 994.287642] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2468782d-9fcb-4171-b02c-5c9001c83535 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.297445] env[61911]: DEBUG oslo_vmware.api [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 994.297445] env[61911]: value = "task-1251491" [ 994.297445] env[61911]: _type = "Task" [ 994.297445] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.306300] env[61911]: DEBUG oslo_vmware.api [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 994.306300] env[61911]: value = "task-1251492" [ 994.306300] env[61911]: _type = "Task" [ 994.306300] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.311939] env[61911]: DEBUG oslo_vmware.api [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251491, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.317108] env[61911]: DEBUG oslo_vmware.api [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251492, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.318193] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.340740] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251489, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.173281} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.341013] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.341862] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb6219c-134a-406e-9ec0-46deacc76250 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.365893] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] b594227d-1647-4e43-8c90-962a9e765388/b594227d-1647-4e43-8c90-962a9e765388.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.366202] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a88cbfc5-40cf-47e2-9e76-a867f1c12e7e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.389026] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 994.389026] env[61911]: value = "task-1251493" [ 994.389026] env[61911]: _type = "Task" [ 994.389026] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.397630] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251493, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.441654] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "f63fa31b-8840-4f95-9514-51f7d5393b25" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.441946] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.466455] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "refresh_cache-f44c597d-ffb8-446b-b814-65733d077dbf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.466904] env[61911]: DEBUG nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Instance network_info: |[{"id": "c2701d58-2bb9-447b-8b02-6f3c3e4cb5de", "address": "fa:16:3e:92:92:c0", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2701d58-2b", "ovs_interfaceid": "c2701d58-2bb9-447b-8b02-6f3c3e4cb5de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 994.467379] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:92:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2701d58-2bb9-447b-8b02-6f3c3e4cb5de', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.476056] env[61911]: DEBUG oslo.service.loopingcall [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.476780] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.477049] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55fea6eb-b679-4b11-ade4-92496e062145 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.502272] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251486, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.058686} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.503505] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 632e811e-b253-47ed-9d1e-6b2538ff804b/632e811e-b253-47ed-9d1e-6b2538ff804b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 994.503786] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 994.504022] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.504022] env[61911]: value = "task-1251494" [ 994.504022] env[61911]: _type = "Task" [ 994.504022] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.504229] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1383788b-9acc-4b8f-9088-ef9215517904 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.514246] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251494, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.515691] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 994.515691] env[61911]: value = "task-1251495" [ 994.515691] env[61911]: _type = "Task" [ 994.515691] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.533562] env[61911]: DEBUG nova.compute.manager [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Received event network-changed-c2701d58-2bb9-447b-8b02-6f3c3e4cb5de {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 994.533796] env[61911]: DEBUG nova.compute.manager [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Refreshing instance network info cache due to event network-changed-c2701d58-2bb9-447b-8b02-6f3c3e4cb5de. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 994.534118] env[61911]: DEBUG oslo_concurrency.lockutils [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] Acquiring lock "refresh_cache-f44c597d-ffb8-446b-b814-65733d077dbf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.534250] env[61911]: DEBUG oslo_concurrency.lockutils [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] Acquired lock "refresh_cache-f44c597d-ffb8-446b-b814-65733d077dbf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.534479] env[61911]: DEBUG nova.network.neutron [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Refreshing network info cache for port c2701d58-2bb9-447b-8b02-6f3c3e4cb5de {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.589339] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 994.589339] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.791s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.589872] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.272s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.591346] env[61911]: INFO nova.compute.claims [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.810663] env[61911]: DEBUG oslo_vmware.api [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251491, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291062} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.814065] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.814287] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 994.814490] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 994.814681] env[61911]: INFO nova.compute.manager [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Took 1.16 seconds to destroy the instance on the hypervisor. [ 994.814957] env[61911]: DEBUG oslo.service.loopingcall [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.815182] env[61911]: DEBUG nova.compute.manager [-] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 994.815279] env[61911]: DEBUG nova.network.neutron [-] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 994.822142] env[61911]: DEBUG oslo_vmware.api [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251492, 'name': ReconfigVM_Task, 'duration_secs': 0.190879} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.822537] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269631', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'name': 'volume-8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c', 'attached_at': '', 'detached_at': '', 'volume_id': '8ade23e8-4f87-4c13-9e03-db70f4c4988d', 'serial': '8ade23e8-4f87-4c13-9e03-db70f4c4988d'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 994.900984] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251493, 'name': ReconfigVM_Task, 'duration_secs': 0.290958} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.901336] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Reconfigured VM instance instance-0000005e to attach disk [datastore1] b594227d-1647-4e43-8c90-962a9e765388/b594227d-1647-4e43-8c90-962a9e765388.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.902079] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2d0c2db-a760-447e-8d1f-4f52d6c52ef0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.909154] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 994.909154] env[61911]: value = "task-1251496" [ 994.909154] env[61911]: _type = "Task" [ 994.909154] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.917935] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251496, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.945353] env[61911]: DEBUG nova.compute.utils [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.019612] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251494, 'name': CreateVM_Task, 'duration_secs': 0.426291} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.023838] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.024838] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.024838] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.025157] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.026289] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14bbb5ac-3158-486f-8415-230ecd0997ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.032563] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251495, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07802} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.034068] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.034479] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 995.034479] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c2ef8d-095f-8eef-0695-d651b1eefd66" [ 995.034479] env[61911]: _type = "Task" [ 995.034479] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.035232] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad35883-285d-4b3b-878a-bf7cbd1f1a66 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.069217] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 632e811e-b253-47ed-9d1e-6b2538ff804b/632e811e-b253-47ed-9d1e-6b2538ff804b.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.069577] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c2ef8d-095f-8eef-0695-d651b1eefd66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.070782] env[61911]: DEBUG nova.network.neutron [-] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.072280] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-160b7e42-4ff7-4c47-986b-49486764c0ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.099122] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 995.099122] env[61911]: value = "task-1251497" [ 995.099122] env[61911]: _type = "Task" [ 995.099122] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.110325] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251497, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.272200] env[61911]: DEBUG nova.network.neutron [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Updated VIF entry in instance network info cache for port c2701d58-2bb9-447b-8b02-6f3c3e4cb5de. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.272680] env[61911]: DEBUG nova.network.neutron [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Updating instance_info_cache with network_info: [{"id": "c2701d58-2bb9-447b-8b02-6f3c3e4cb5de", "address": "fa:16:3e:92:92:c0", "network": {"id": "8574ed68-d2f1-44ff-af3d-65cbc72d1323", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-646638887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5be8ec1d59a14d618906ec32d0e134e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2701d58-2b", "ovs_interfaceid": "c2701d58-2bb9-447b-8b02-6f3c3e4cb5de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.368283] env[61911]: DEBUG nova.objects.instance [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lazy-loading 'flavor' on Instance uuid 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.420274] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251496, 'name': Rename_Task, 'duration_secs': 0.15524} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.420567] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 995.420814] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47720b4e-d546-495e-9ade-d16e97d59523 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.427540] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 995.427540] env[61911]: value = "task-1251498" [ 995.427540] env[61911]: _type = "Task" [ 995.427540] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.435103] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.447905] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.549141] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c2ef8d-095f-8eef-0695-d651b1eefd66, 'name': SearchDatastore_Task, 'duration_secs': 0.027938} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.549467] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.549718] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.549956] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.550127] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.550312] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.550609] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f37beb9b-5369-4cec-894b-8d0f653e2475 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.559984] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.560177] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.560916] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d380670c-2e68-4109-97c6-528a03460c3d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.566078] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 995.566078] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bbbaee-4ad2-5270-4df0-f0eed0fa5439" [ 995.566078] env[61911]: _type = "Task" [ 995.566078] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.573729] env[61911]: INFO nova.compute.manager [-] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Took 0.76 seconds to deallocate network for instance. [ 995.574306] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bbbaee-4ad2-5270-4df0-f0eed0fa5439, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.612265] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251497, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.763702] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d124dc6-0aa8-43df-8d48-acc28f98f2c8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.771518] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db48f60-c40a-4c37-97d6-13bf7fb2fa5e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.775121] env[61911]: DEBUG oslo_concurrency.lockutils [req-d4c733ea-254e-4391-b0bc-432220fe1d2f req-1bfb64f8-04f9-4ae3-9656-eeb3be3ca2b8 service nova] Releasing lock "refresh_cache-f44c597d-ffb8-446b-b814-65733d077dbf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.802240] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44eb25e-532d-425e-a926-aa9cdcf9b114 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.810094] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526a492c-2582-4403-9880-023647aacb1e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.823592] env[61911]: DEBUG nova.compute.provider_tree [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.938219] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251498, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.077437] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bbbaee-4ad2-5270-4df0-f0eed0fa5439, 'name': SearchDatastore_Task, 'duration_secs': 0.027217} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.078270] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-012c5a6d-88e0-4ad1-9117-38319014cf02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.081478] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.085210] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 996.085210] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d068f4-b13f-f309-6eb4-3fab842473fb" [ 996.085210] env[61911]: _type = "Task" [ 996.085210] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.095048] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d068f4-b13f-f309-6eb4-3fab842473fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.115799] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251497, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.327300] env[61911]: DEBUG nova.scheduler.client.report [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 996.377101] env[61911]: DEBUG oslo_concurrency.lockutils [None req-950dbe32-1fde-44be-a366-17c8f3f9300a tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.749s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.440639] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251498, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.525904] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "f63fa31b-8840-4f95-9514-51f7d5393b25" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.526200] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.526447] env[61911]: INFO nova.compute.manager [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Attaching volume 19d8968b-4a55-4aff-b72f-65e2be6ff7e9 to /dev/sdb [ 996.560158] env[61911]: DEBUG nova.compute.manager [req-0097a3bb-53f6-4dc0-8017-904d1e97cf93 req-66e0c8b1-e2ee-48a8-8606-cdd59a44e737 service nova] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Received event network-vif-deleted-227cefa9-0b37-43fc-9c0b-da58e16ae969 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 996.585665] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd757035-ed33-443c-ad8c-3cdaa6bfab81 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.601508] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d459b5c-20ee-4832-8185-32a09516035f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.606610] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d068f4-b13f-f309-6eb4-3fab842473fb, 'name': SearchDatastore_Task, 'duration_secs': 0.012502} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.607178] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.607613] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] f44c597d-ffb8-446b-b814-65733d077dbf/f44c597d-ffb8-446b-b814-65733d077dbf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.610842] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec163233-2d50-46b4-a847-9615965e75e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.621449] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251497, 'name': ReconfigVM_Task, 'duration_secs': 1.068874} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.626986] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 632e811e-b253-47ed-9d1e-6b2538ff804b/632e811e-b253-47ed-9d1e-6b2538ff804b.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.628431] env[61911]: DEBUG nova.virt.block_device [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Updating existing volume attachment record: aa4589fd-acd3-4452-a49b-001e04107cc2 {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 996.630823] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 996.630823] env[61911]: value = "task-1251499" [ 996.630823] env[61911]: _type = "Task" [ 996.630823] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.631292] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40547f2f-3001-4ac5-8ebd-7c8edf5f6b02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.641673] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251499, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.642806] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 996.642806] env[61911]: value = "task-1251500" [ 996.642806] env[61911]: _type = "Task" [ 996.642806] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.655281] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251500, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.833039] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.833610] env[61911]: DEBUG nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 996.836276] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.755s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.836500] env[61911]: DEBUG nova.objects.instance [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lazy-loading 'resources' on Instance uuid d9234a5f-4d26-4bdc-8f32-3120830a8abe {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.941285] env[61911]: DEBUG oslo_vmware.api [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251498, 'name': PowerOnVM_Task, 'duration_secs': 1.039901} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.941719] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 996.941885] env[61911]: INFO nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Took 10.39 seconds to spawn the instance on the hypervisor. [ 996.942112] env[61911]: DEBUG nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 996.943385] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ac9077-7422-47c6-8e2a-bcb4d4005b39 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.108549] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.108953] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.109194] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.109442] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.109671] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.112259] env[61911]: INFO nova.compute.manager [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Terminating instance [ 997.143032] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251499, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508636} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.143404] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] f44c597d-ffb8-446b-b814-65733d077dbf/f44c597d-ffb8-446b-b814-65733d077dbf.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.143685] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.146668] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa73521f-8157-41da-9195-6c57dbeff462 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.154060] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251500, 'name': Rename_Task, 'duration_secs': 0.169668} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.155315] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.155618] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 997.155618] env[61911]: value = "task-1251504" [ 997.155618] env[61911]: _type = "Task" [ 997.155618] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.155806] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-088d2e19-6807-40d1-b58f-41083eba53b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.164965] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251504, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.166170] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 997.166170] env[61911]: value = "task-1251505" [ 997.166170] env[61911]: _type = "Task" [ 997.166170] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.173279] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.339145] env[61911]: DEBUG nova.compute.utils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.343432] env[61911]: DEBUG nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 997.343693] env[61911]: DEBUG nova.network.neutron [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 997.383565] env[61911]: DEBUG nova.policy [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe50c877fcef489cb798fafe73caaf8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813b3181120245e594809d096a621675', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 997.469309] env[61911]: INFO nova.compute.manager [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Took 19.09 seconds to build instance. [ 997.515121] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf4988b-5bab-4b05-9615-c3e424d179b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.525212] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6caea223-9409-4281-9258-065c4ef40bef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.555974] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e6094e6-2666-4080-b8a3-d78276776341 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.564273] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3a2ce1-584a-4124-85a5-a6ae74f2637b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.579188] env[61911]: DEBUG nova.compute.provider_tree [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.616813] env[61911]: DEBUG nova.compute.manager [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 997.617136] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.618015] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa4c19a-5942-450a-9b19-9a039bbd5610 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.626699] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.626960] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca2d6f84-d8c0-4ad5-a61f-637ab6f46532 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.634228] env[61911]: DEBUG oslo_vmware.api [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 997.634228] env[61911]: value = "task-1251506" [ 997.634228] env[61911]: _type = "Task" [ 997.634228] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.642613] env[61911]: DEBUG oslo_vmware.api [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251506, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.656137] env[61911]: DEBUG nova.network.neutron [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Successfully created port: 73e4548b-9760-484e-9023-f467279085d6 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.667431] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251504, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.676570] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.844374] env[61911]: DEBUG nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 997.971944] env[61911]: DEBUG oslo_concurrency.lockutils [None req-975752b0-4269-4ee4-9b7f-cbb633767c96 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "b594227d-1647-4e43-8c90-962a9e765388" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.605s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.082810] env[61911]: DEBUG nova.scheduler.client.report [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 998.145860] env[61911]: DEBUG oslo_vmware.api [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251506, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.169066] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251504, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.889025} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.172950] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.174112] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b47fc96-4be6-4222-b685-d3990ec61746 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.183703] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.207072] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] f44c597d-ffb8-446b-b814-65733d077dbf/f44c597d-ffb8-446b-b814-65733d077dbf.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.207072] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9f35691-a9c9-4cf5-b1c4-e57c4e228816 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.228056] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 998.228056] env[61911]: value = "task-1251507" [ 998.228056] env[61911]: _type = "Task" [ 998.228056] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.238299] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251507, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.588693] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.751s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.615009] env[61911]: INFO nova.scheduler.client.report [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleted allocations for instance d9234a5f-4d26-4bdc-8f32-3120830a8abe [ 998.646801] env[61911]: DEBUG oslo_vmware.api [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251506, 'name': PowerOffVM_Task, 'duration_secs': 0.654416} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.647163] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.647402] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.647732] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-762724de-249a-47c7-a4dd-3a543d77b78c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.680828] env[61911]: DEBUG oslo_vmware.api [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251505, 'name': PowerOnVM_Task, 'duration_secs': 1.496313} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.681074] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.681269] env[61911]: INFO nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Took 9.29 seconds to spawn the instance on the hypervisor. [ 998.681508] env[61911]: DEBUG nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 998.682699] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd6ccdf-c65e-44ec-8cba-1c292d261f86 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.743039] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251507, 'name': ReconfigVM_Task, 'duration_secs': 0.472306} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.743039] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Reconfigured VM instance instance-00000060 to attach disk [datastore2] f44c597d-ffb8-446b-b814-65733d077dbf/f44c597d-ffb8-446b-b814-65733d077dbf.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.743449] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05dc32b7-bfbb-4079-a7e6-6880961a3d75 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.750713] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 998.750713] env[61911]: value = "task-1251509" [ 998.750713] env[61911]: _type = "Task" [ 998.750713] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.759677] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251509, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.770080] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.770311] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.770517] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleting the datastore file [datastore1] 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.770790] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c983879c-2c27-4422-bd96-d6423d7ce89a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.777416] env[61911]: DEBUG oslo_vmware.api [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 998.777416] env[61911]: value = "task-1251510" [ 998.777416] env[61911]: _type = "Task" [ 998.777416] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.785711] env[61911]: DEBUG oslo_vmware.api [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251510, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.854517] env[61911]: DEBUG nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 998.880814] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 998.881149] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.881355] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 998.881575] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.881749] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 998.881917] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 998.882150] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 998.882451] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 998.882702] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 998.882887] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 998.883080] env[61911]: DEBUG nova.virt.hardware [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 998.883941] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b31851-eabd-4a16-a917-7316a3abba55 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.891492] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1adf3e-1ce6-40bb-8882-8df7ddec61af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.038703] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "fddba36d-1b15-43fb-9e99-68880e8a235c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.038703] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.038703] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "fddba36d-1b15-43fb-9e99-68880e8a235c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.041079] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.041079] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.042361] env[61911]: INFO nova.compute.manager [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Terminating instance [ 999.049189] env[61911]: DEBUG nova.compute.manager [req-bce5f140-cde2-4d55-bc5c-04726c1b3849 req-660e80ec-52a2-4790-b50a-04c987ad7a01 service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Received event network-vif-plugged-73e4548b-9760-484e-9023-f467279085d6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 999.049416] env[61911]: DEBUG oslo_concurrency.lockutils [req-bce5f140-cde2-4d55-bc5c-04726c1b3849 req-660e80ec-52a2-4790-b50a-04c987ad7a01 service nova] Acquiring lock "a03413fe-e74d-42e4-83ea-a9c19318526e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.049647] env[61911]: DEBUG oslo_concurrency.lockutils [req-bce5f140-cde2-4d55-bc5c-04726c1b3849 req-660e80ec-52a2-4790-b50a-04c987ad7a01 service nova] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.049831] env[61911]: DEBUG oslo_concurrency.lockutils [req-bce5f140-cde2-4d55-bc5c-04726c1b3849 req-660e80ec-52a2-4790-b50a-04c987ad7a01 service nova] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.050014] env[61911]: DEBUG nova.compute.manager [req-bce5f140-cde2-4d55-bc5c-04726c1b3849 req-660e80ec-52a2-4790-b50a-04c987ad7a01 service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] No waiting events found dispatching network-vif-plugged-73e4548b-9760-484e-9023-f467279085d6 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 999.050193] env[61911]: WARNING nova.compute.manager [req-bce5f140-cde2-4d55-bc5c-04726c1b3849 req-660e80ec-52a2-4790-b50a-04c987ad7a01 service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Received unexpected event network-vif-plugged-73e4548b-9760-484e-9023-f467279085d6 for instance with vm_state building and task_state spawning. [ 999.123825] env[61911]: DEBUG oslo_concurrency.lockutils [None req-823c35d0-20ba-4e8f-a665-c647af3861c5 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "d9234a5f-4d26-4bdc-8f32-3120830a8abe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.973s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.199886] env[61911]: INFO nova.compute.manager [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Took 20.58 seconds to build instance. [ 999.262145] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251509, 'name': Rename_Task, 'duration_secs': 0.203179} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.262145] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 999.262145] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc30b989-b8bb-4807-98b4-40d776cc7076 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.268378] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 999.268378] env[61911]: value = "task-1251512" [ 999.268378] env[61911]: _type = "Task" [ 999.268378] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.277458] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251512, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.284693] env[61911]: DEBUG oslo_vmware.api [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251510, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213066} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.284982] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.285233] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.285511] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.285755] env[61911]: INFO nova.compute.manager [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Took 1.67 seconds to destroy the instance on the hypervisor. [ 999.286060] env[61911]: DEBUG oslo.service.loopingcall [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.286312] env[61911]: DEBUG nova.compute.manager [-] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 999.286417] env[61911]: DEBUG nova.network.neutron [-] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.546834] env[61911]: DEBUG nova.compute.manager [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 999.548800] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.549376] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4c5e22-57f7-41f5-b928-4f9fa6f98d8e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.558186] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.558497] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c261813d-7c3f-4b71-a2ee-bd6bc9e52505 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.564850] env[61911]: DEBUG oslo_vmware.api [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 999.564850] env[61911]: value = "task-1251513" [ 999.564850] env[61911]: _type = "Task" [ 999.564850] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.572508] env[61911]: DEBUG oslo_vmware.api [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251513, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.612363] env[61911]: DEBUG nova.network.neutron [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Successfully updated port: 73e4548b-9760-484e-9023-f467279085d6 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.702102] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9af6b1f3-03e5-42bb-b978-1d3c8e6f4ca2 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.090s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.708264] env[61911]: DEBUG nova.compute.manager [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Received event network-changed-73e4548b-9760-484e-9023-f467279085d6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 999.708516] env[61911]: DEBUG nova.compute.manager [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Refreshing instance network info cache due to event network-changed-73e4548b-9760-484e-9023-f467279085d6. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 999.708852] env[61911]: DEBUG oslo_concurrency.lockutils [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] Acquiring lock "refresh_cache-a03413fe-e74d-42e4-83ea-a9c19318526e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.708960] env[61911]: DEBUG oslo_concurrency.lockutils [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] Acquired lock "refresh_cache-a03413fe-e74d-42e4-83ea-a9c19318526e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.709322] env[61911]: DEBUG nova.network.neutron [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Refreshing network info cache for port 73e4548b-9760-484e-9023-f467279085d6 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.779809] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251512, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.870092] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.870376] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.870606] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.870828] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.870997] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.873321] env[61911]: INFO nova.compute.manager [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Terminating instance [ 1000.075452] env[61911]: DEBUG oslo_vmware.api [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251513, 'name': PowerOffVM_Task, 'duration_secs': 0.205819} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.075750] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.075924] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.076219] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eaab2c81-b5ae-4c71-ab4c-e8980c90683b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.110232] env[61911]: DEBUG nova.compute.manager [req-51b9dbe9-c192-44ab-9e80-fdb0596ba192 req-f178bd4b-11a8-4951-8b09-9a3a95fc1777 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Received event network-vif-deleted-7ce08c90-28e1-4989-9616-05bc8d5ee616 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1000.110449] env[61911]: INFO nova.compute.manager [req-51b9dbe9-c192-44ab-9e80-fdb0596ba192 req-f178bd4b-11a8-4951-8b09-9a3a95fc1777 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Neutron deleted interface 7ce08c90-28e1-4989-9616-05bc8d5ee616; detaching it from the instance and deleting it from the info cache [ 1000.110649] env[61911]: DEBUG nova.network.neutron [req-51b9dbe9-c192-44ab-9e80-fdb0596ba192 req-f178bd4b-11a8-4951-8b09-9a3a95fc1777 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.115911] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "refresh_cache-a03413fe-e74d-42e4-83ea-a9c19318526e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.140908] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.141159] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.141585] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleting the datastore file [datastore2] fddba36d-1b15-43fb-9e99-68880e8a235c {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.142292] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6af6a70-3e26-4c94-b207-04eadf7fdd84 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.149704] env[61911]: DEBUG oslo_vmware.api [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1000.149704] env[61911]: value = "task-1251515" [ 1000.149704] env[61911]: _type = "Task" [ 1000.149704] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.158173] env[61911]: DEBUG oslo_vmware.api [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251515, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.245937] env[61911]: DEBUG nova.network.neutron [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.281747] env[61911]: DEBUG oslo_vmware.api [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251512, 'name': PowerOnVM_Task, 'duration_secs': 0.946307} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.282036] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.282259] env[61911]: INFO nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Took 8.34 seconds to spawn the instance on the hypervisor. [ 1000.282435] env[61911]: DEBUG nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1000.283321] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c2ac66-92b8-48e1-ac2a-5dfe9556b151 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.333247] env[61911]: DEBUG nova.network.neutron [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.377593] env[61911]: DEBUG nova.compute.manager [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1000.377770] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.378775] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a5a16b-402c-4e14-967d-634f52df335d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.386365] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.386614] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ad1775c-f8dd-41c0-b3b6-8c668052d752 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.392129] env[61911]: DEBUG oslo_vmware.api [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 1000.392129] env[61911]: value = "task-1251516" [ 1000.392129] env[61911]: _type = "Task" [ 1000.392129] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.400496] env[61911]: DEBUG oslo_vmware.api [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.585194] env[61911]: DEBUG nova.network.neutron [-] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.612925] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb5d6918-68ad-441e-9055-d2d4f7307673 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.623036] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360f462a-4c8b-4912-95f4-340761388b93 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.652080] env[61911]: DEBUG nova.compute.manager [req-51b9dbe9-c192-44ab-9e80-fdb0596ba192 req-f178bd4b-11a8-4951-8b09-9a3a95fc1777 service nova] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Detach interface failed, port_id=7ce08c90-28e1-4989-9616-05bc8d5ee616, reason: Instance 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1000.660114] env[61911]: DEBUG oslo_vmware.api [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251515, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146111} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.660389] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.660582] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.660781] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.660967] env[61911]: INFO nova.compute.manager [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1000.661238] env[61911]: DEBUG oslo.service.loopingcall [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.661433] env[61911]: DEBUG nova.compute.manager [-] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1000.661531] env[61911]: DEBUG nova.network.neutron [-] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.805245] env[61911]: INFO nova.compute.manager [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Took 14.90 seconds to build instance. [ 1000.836205] env[61911]: DEBUG oslo_concurrency.lockutils [req-095d2f21-6fec-4479-a53a-fbf1ade54310 req-918a1103-b918-412c-a373-dd77d521eb1b service nova] Releasing lock "refresh_cache-a03413fe-e74d-42e4-83ea-a9c19318526e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.836624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "refresh_cache-a03413fe-e74d-42e4-83ea-a9c19318526e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.836817] env[61911]: DEBUG nova.network.neutron [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.902636] env[61911]: DEBUG oslo_vmware.api [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251516, 'name': PowerOffVM_Task, 'duration_secs': 0.180035} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.903095] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.903335] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.903630] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad134297-998b-4a3f-b439-911c13eb7903 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.968672] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.969393] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.969655] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleting the datastore file [datastore1] b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.969943] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa1714e2-ecd9-498b-82ac-d352c4e4bbb5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.977936] env[61911]: DEBUG oslo_vmware.api [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 1000.977936] env[61911]: value = "task-1251518" [ 1000.977936] env[61911]: _type = "Task" [ 1000.977936] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.986311] env[61911]: DEBUG oslo_vmware.api [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.088323] env[61911]: INFO nova.compute.manager [-] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Took 1.80 seconds to deallocate network for instance. [ 1001.184229] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1001.184547] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269659', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'name': 'volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f63fa31b-8840-4f95-9514-51f7d5393b25', 'attached_at': '', 'detached_at': '', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'serial': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1001.185530] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44081c77-17ff-417c-8514-fa6bc4d457b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.204551] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4aab83-9776-492e-890b-1f985cadd313 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.234726] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9/volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.235181] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e4455ab-245e-48be-bb91-a76b99bf646b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.255769] env[61911]: DEBUG oslo_vmware.api [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1001.255769] env[61911]: value = "task-1251519" [ 1001.255769] env[61911]: _type = "Task" [ 1001.255769] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.265580] env[61911]: DEBUG oslo_vmware.api [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251519, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.304393] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b16d2a71-6018-4a3d-8072-eeed2b1d6339 tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f44c597d-ffb8-446b-b814-65733d077dbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.412s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.368894] env[61911]: DEBUG nova.network.neutron [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.488353] env[61911]: DEBUG oslo_vmware.api [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283617} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.488667] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.488865] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.489142] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.489383] env[61911]: INFO nova.compute.manager [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1001.489687] env[61911]: DEBUG oslo.service.loopingcall [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.489921] env[61911]: DEBUG nova.compute.manager [-] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1001.490059] env[61911]: DEBUG nova.network.neutron [-] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.494373] env[61911]: DEBUG nova.compute.manager [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1001.495220] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e2034b-a7e5-46ef-9c36-e6e6f5fa2860 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.581902] env[61911]: DEBUG nova.network.neutron [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Updating instance_info_cache with network_info: [{"id": "73e4548b-9760-484e-9023-f467279085d6", "address": "fa:16:3e:ae:fc:76", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73e4548b-97", "ovs_interfaceid": "73e4548b-9760-484e-9023-f467279085d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.594783] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.595159] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.595406] env[61911]: DEBUG nova.objects.instance [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lazy-loading 'resources' on Instance uuid 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.766769] env[61911]: DEBUG oslo_vmware.api [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251519, 'name': ReconfigVM_Task, 'duration_secs': 0.371655} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.767024] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Reconfigured VM instance instance-0000005d to attach disk [datastore1] volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9/volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.774320] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e8b266d-edfd-4dd4-b08d-d65f23e9b41e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.786809] env[61911]: DEBUG nova.compute.manager [req-c941f31b-1fd9-4ba2-9e37-7d724c60022c req-504d0749-931a-48b3-a958-e845d9d6bd77 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Received event network-vif-deleted-022b3630-95b4-40fa-8930-c97b6b008fd3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1001.787450] env[61911]: INFO nova.compute.manager [req-c941f31b-1fd9-4ba2-9e37-7d724c60022c req-504d0749-931a-48b3-a958-e845d9d6bd77 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Neutron deleted interface 022b3630-95b4-40fa-8930-c97b6b008fd3; detaching it from the instance and deleting it from the info cache [ 1001.787450] env[61911]: DEBUG nova.network.neutron [req-c941f31b-1fd9-4ba2-9e37-7d724c60022c req-504d0749-931a-48b3-a958-e845d9d6bd77 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.796021] env[61911]: DEBUG oslo_vmware.api [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1001.796021] env[61911]: value = "task-1251520" [ 1001.796021] env[61911]: _type = "Task" [ 1001.796021] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.805507] env[61911]: DEBUG oslo_vmware.api [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251520, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.816710] env[61911]: DEBUG nova.network.neutron [-] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.006952] env[61911]: INFO nova.compute.manager [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] instance snapshotting [ 1002.010835] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce44684-21ad-40c2-898e-375d505343e0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.029715] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4dc8b84-f601-4e82-9937-bdf7698cd72f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.085213] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "refresh_cache-a03413fe-e74d-42e4-83ea-a9c19318526e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.085560] env[61911]: DEBUG nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Instance network_info: |[{"id": "73e4548b-9760-484e-9023-f467279085d6", "address": "fa:16:3e:ae:fc:76", "network": {"id": "8d571001-9263-44e9-8997-3399ab9232bd", "bridge": "br-int", "label": "tempest-ServersTestJSON-819944901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "813b3181120245e594809d096a621675", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73e4548b-97", "ovs_interfaceid": "73e4548b-9760-484e-9023-f467279085d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1002.086011] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:fc:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73e4548b-9760-484e-9023-f467279085d6', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.093605] env[61911]: DEBUG oslo.service.loopingcall [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.094755] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.095009] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-806f8edb-b412-4cbc-be39-8e5e42a28887 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.117850] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.117850] env[61911]: value = "task-1251521" [ 1002.117850] env[61911]: _type = "Task" [ 1002.117850] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.127693] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251521, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.136411] env[61911]: DEBUG nova.compute.manager [req-de9a4f2a-77f7-4063-a9a0-79df2c090312 req-a3ab05fb-f249-42fb-a147-705352725b8d service nova] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Received event network-vif-deleted-33e8e080-830e-4a6a-8e9b-39eded7bf6de {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1002.249598] env[61911]: DEBUG nova.network.neutron [-] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.290203] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42a549ea-90ca-42c9-9a4c-cac90e2676c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.302159] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c25c9b3-de82-4d13-93ea-bb4881a927e7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.313228] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082bdbac-659e-49f2-b468-5bf8209c27e5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.320433] env[61911]: INFO nova.compute.manager [-] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Took 1.66 seconds to deallocate network for instance. [ 1002.320771] env[61911]: DEBUG oslo_vmware.api [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251520, 'name': ReconfigVM_Task, 'duration_secs': 0.138109} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.324011] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269659', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'name': 'volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f63fa31b-8840-4f95-9514-51f7d5393b25', 'attached_at': '', 'detached_at': '', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'serial': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1002.338486] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495dffc4-a586-4643-9dff-1dabcdb77757 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.342156] env[61911]: DEBUG nova.compute.manager [req-c941f31b-1fd9-4ba2-9e37-7d724c60022c req-504d0749-931a-48b3-a958-e845d9d6bd77 service nova] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Detach interface failed, port_id=022b3630-95b4-40fa-8930-c97b6b008fd3, reason: Instance b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1002.375659] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3714cabe-3ec7-49a5-8e37-89df6aaa8faa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.384652] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d50ca7-6ed3-4532-858d-a9df856dffb3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.400255] env[61911]: DEBUG nova.compute.provider_tree [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.541072] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1002.541072] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-59c4b752-3d43-44de-83e5-7abd743161be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.548249] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1002.548249] env[61911]: value = "task-1251522" [ 1002.548249] env[61911]: _type = "Task" [ 1002.548249] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.556330] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251522, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.561863] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f44c597d-ffb8-446b-b814-65733d077dbf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.562117] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f44c597d-ffb8-446b-b814-65733d077dbf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.562335] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "f44c597d-ffb8-446b-b814-65733d077dbf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.562525] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f44c597d-ffb8-446b-b814-65733d077dbf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.562775] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f44c597d-ffb8-446b-b814-65733d077dbf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.565233] env[61911]: INFO nova.compute.manager [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Terminating instance [ 1002.631161] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251521, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.753017] env[61911]: INFO nova.compute.manager [-] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Took 1.26 seconds to deallocate network for instance. [ 1002.846572] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.880608] env[61911]: DEBUG nova.objects.instance [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'flavor' on Instance uuid f63fa31b-8840-4f95-9514-51f7d5393b25 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.902255] env[61911]: DEBUG nova.scheduler.client.report [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1003.060549] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251522, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.069874] env[61911]: DEBUG nova.compute.manager [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1003.070225] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.071653] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c127f7c1-c0b3-4a15-8029-06f2b5a1a41f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.079432] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.079695] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f42224f-51ad-4953-bfa9-2cdf39180340 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.086405] env[61911]: DEBUG oslo_vmware.api [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 1003.086405] env[61911]: value = "task-1251523" [ 1003.086405] env[61911]: _type = "Task" [ 1003.086405] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.094214] env[61911]: DEBUG oslo_vmware.api [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251523, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.128172] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251521, 'name': CreateVM_Task, 'duration_secs': 0.564545} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.128335] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.128976] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.129168] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.129499] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.129757] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb4b59c4-81e7-4855-8f25-0672391622cd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.134261] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1003.134261] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b80d6d-c3a5-2840-4a8b-a92c9484de4e" [ 1003.134261] env[61911]: _type = "Task" [ 1003.134261] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.141667] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b80d6d-c3a5-2840-4a8b-a92c9484de4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.260286] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.387027] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6039b907-08e9-4408-8ed9-9f12e5ed77ff tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 6.860s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.407453] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.812s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.409680] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.563s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.409919] env[61911]: DEBUG nova.objects.instance [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'resources' on Instance uuid fddba36d-1b15-43fb-9e99-68880e8a235c {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.434564] env[61911]: INFO nova.scheduler.client.report [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleted allocations for instance 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c [ 1003.559064] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251522, 'name': CreateSnapshot_Task, 'duration_secs': 0.547648} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.559358] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1003.560104] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bbed9f-37e5-4df7-8999-703afa625ebe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.595556] env[61911]: DEBUG oslo_vmware.api [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251523, 'name': PowerOffVM_Task, 'duration_secs': 0.266707} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.595835] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.596022] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.596274] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f340e60-264e-4fbe-9812-af2ca1704ba1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.644518] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b80d6d-c3a5-2840-4a8b-a92c9484de4e, 'name': SearchDatastore_Task, 'duration_secs': 0.009725} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.644840] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.645109] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.645364] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.645520] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.645705] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.645969] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d85c655-d7ad-45c5-ab9d-85ae6ca425cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.654589] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.654817] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.655031] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleting the datastore file [datastore2] f44c597d-ffb8-446b-b814-65733d077dbf {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.656073] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-282218b8-26bb-48b0-bf7b-c4e3169c07f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.657799] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.657977] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.658665] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83bde3e1-d92f-46e6-b939-99e5dbbef942 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.664118] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1003.664118] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5250bc88-f5f2-8772-f6ba-0a1db3f035bd" [ 1003.664118] env[61911]: _type = "Task" [ 1003.664118] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.668213] env[61911]: DEBUG oslo_vmware.api [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for the task: (returnval){ [ 1003.668213] env[61911]: value = "task-1251525" [ 1003.668213] env[61911]: _type = "Task" [ 1003.668213] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.674108] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5250bc88-f5f2-8772-f6ba-0a1db3f035bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.678229] env[61911]: DEBUG oslo_vmware.api [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251525, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.906567] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "f63fa31b-8840-4f95-9514-51f7d5393b25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.906567] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.906567] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "f63fa31b-8840-4f95-9514-51f7d5393b25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.906567] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.906567] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.908714] env[61911]: INFO nova.compute.manager [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Terminating instance [ 1003.944674] env[61911]: DEBUG oslo_concurrency.lockutils [None req-031f5464-6931-49b8-9a25-c1bac005ba6b tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.836s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.052168] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02cff8f7-48e4-4d91-821a-8f289e9b87a3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.058762] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d48f5e0-fbdc-4f3a-95fc-bd7ed4cd8977 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.094894] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1004.095789] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-445efbc5-d7ee-4a96-9e1f-6a976b9d4b7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.099292] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c2efb4-2a0e-47be-a807-1ead914b4568 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.108311] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5687d943-be60-4fc2-86ac-fc0481eea38a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.112148] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1004.112148] env[61911]: value = "task-1251526" [ 1004.112148] env[61911]: _type = "Task" [ 1004.112148] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.122909] env[61911]: DEBUG nova.compute.provider_tree [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.128945] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251526, 'name': CloneVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.180045] env[61911]: DEBUG oslo_vmware.api [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Task: {'id': task-1251525, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150742} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.180275] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5250bc88-f5f2-8772-f6ba-0a1db3f035bd, 'name': SearchDatastore_Task, 'duration_secs': 0.011647} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.180520] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.180759] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.180988] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.181214] env[61911]: INFO nova.compute.manager [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1004.181480] env[61911]: DEBUG oslo.service.loopingcall [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.182264] env[61911]: DEBUG nova.compute.manager [-] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1004.182372] env[61911]: DEBUG nova.network.neutron [-] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.184044] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ac81fda-d4f5-44ea-87c8-95d475c52ec8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.189172] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1004.189172] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c0398e-787f-59c6-9249-dbf8d3e79415" [ 1004.189172] env[61911]: _type = "Task" [ 1004.189172] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.197154] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c0398e-787f-59c6-9249-dbf8d3e79415, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.415424] env[61911]: DEBUG nova.compute.manager [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1004.415424] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.415424] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80b0e972-c1b4-453f-a279-328ec847d369 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.421805] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1004.421805] env[61911]: value = "task-1251527" [ 1004.421805] env[61911]: _type = "Task" [ 1004.421805] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.431246] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.508422] env[61911]: DEBUG nova.compute.manager [req-e5d7c73d-4c46-467b-ad6a-449327850822 req-98c99e16-2e2d-4a32-beeb-5af07286bec3 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Received event network-vif-deleted-c2701d58-2bb9-447b-8b02-6f3c3e4cb5de {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1004.508546] env[61911]: INFO nova.compute.manager [req-e5d7c73d-4c46-467b-ad6a-449327850822 req-98c99e16-2e2d-4a32-beeb-5af07286bec3 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Neutron deleted interface c2701d58-2bb9-447b-8b02-6f3c3e4cb5de; detaching it from the instance and deleting it from the info cache [ 1004.508831] env[61911]: DEBUG nova.network.neutron [req-e5d7c73d-4c46-467b-ad6a-449327850822 req-98c99e16-2e2d-4a32-beeb-5af07286bec3 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.622155] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251526, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.625539] env[61911]: DEBUG nova.scheduler.client.report [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1004.699997] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c0398e-787f-59c6-9249-dbf8d3e79415, 'name': SearchDatastore_Task, 'duration_secs': 0.012291} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.700333] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.700571] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] a03413fe-e74d-42e4-83ea-a9c19318526e/a03413fe-e74d-42e4-83ea-a9c19318526e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.700840] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45ad517b-6a77-42d7-89cc-b6a1f3e7d502 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.707234] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1004.707234] env[61911]: value = "task-1251528" [ 1004.707234] env[61911]: _type = "Task" [ 1004.707234] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.715400] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251528, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.933414] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251527, 'name': PowerOffVM_Task, 'duration_secs': 0.219694} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.933747] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.934039] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1004.934260] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269659', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'name': 'volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f63fa31b-8840-4f95-9514-51f7d5393b25', 'attached_at': '', 'detached_at': '', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'serial': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1004.935073] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81a9b98-f67c-42ce-bf99-14f5ca62a637 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.956177] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d990645-9130-4e66-a057-06a9b6c4a8c1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.964385] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedfb09a-24c5-45ef-8011-d3d99ea92429 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.985922] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938c3e34-f6c5-4308-97ff-5b712f2526b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.988730] env[61911]: DEBUG nova.network.neutron [-] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.004582] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] The volume has not been displaced from its original location: [datastore1] volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9/volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1005.010406] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1005.011841] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82f27095-5aed-4090-866d-09c144d58fdb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.024971] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c218123d-680b-4211-95cb-d267b69faf0f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.035395] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919450c3-9361-4d5a-9b40-fdb8fcba962c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.049075] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1005.049075] env[61911]: value = "task-1251530" [ 1005.049075] env[61911]: _type = "Task" [ 1005.049075] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.059019] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251530, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.068921] env[61911]: DEBUG nova.compute.manager [req-e5d7c73d-4c46-467b-ad6a-449327850822 req-98c99e16-2e2d-4a32-beeb-5af07286bec3 service nova] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Detach interface failed, port_id=c2701d58-2bb9-447b-8b02-6f3c3e4cb5de, reason: Instance f44c597d-ffb8-446b-b814-65733d077dbf could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1005.123423] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251526, 'name': CloneVM_Task} progress is 95%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.131533] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.722s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.134114] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.874s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.134386] env[61911]: DEBUG nova.objects.instance [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lazy-loading 'resources' on Instance uuid b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.161208] env[61911]: INFO nova.scheduler.client.report [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted allocations for instance fddba36d-1b15-43fb-9e99-68880e8a235c [ 1005.218135] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251528, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470568} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.218564] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] a03413fe-e74d-42e4-83ea-a9c19318526e/a03413fe-e74d-42e4-83ea-a9c19318526e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.218829] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.219194] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c24670d-4092-4281-a042-91d9c7b2d5a3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.226457] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1005.226457] env[61911]: value = "task-1251531" [ 1005.226457] env[61911]: _type = "Task" [ 1005.226457] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.235388] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251531, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.493640] env[61911]: INFO nova.compute.manager [-] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Took 1.31 seconds to deallocate network for instance. [ 1005.559404] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251530, 'name': ReconfigVM_Task, 'duration_secs': 0.316381} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.559705] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1005.564320] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f542cab9-a21b-4941-b0a0-d9deb5431258 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.579434] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1005.579434] env[61911]: value = "task-1251532" [ 1005.579434] env[61911]: _type = "Task" [ 1005.579434] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.589175] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251532, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.622993] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251526, 'name': CloneVM_Task, 'duration_secs': 1.287603} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.623323] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Created linked-clone VM from snapshot [ 1005.624073] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5d666f-fb68-43b5-b3d9-9eb88992f598 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.631231] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Uploading image 8cdf5b0c-7e3c-4b1b-8bfc-345854bccdff {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1005.657547] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1005.657547] env[61911]: value = "vm-269662" [ 1005.657547] env[61911]: _type = "VirtualMachine" [ 1005.657547] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1005.657925] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0c43187e-b524-47cc-a536-ee7d44bb973d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.670300] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lease: (returnval){ [ 1005.670300] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52170662-3e1b-030f-0338-0bd11bbac05a" [ 1005.670300] env[61911]: _type = "HttpNfcLease" [ 1005.670300] env[61911]: } obtained for exporting VM: (result){ [ 1005.670300] env[61911]: value = "vm-269662" [ 1005.670300] env[61911]: _type = "VirtualMachine" [ 1005.670300] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1005.670593] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the lease: (returnval){ [ 1005.670593] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52170662-3e1b-030f-0338-0bd11bbac05a" [ 1005.670593] env[61911]: _type = "HttpNfcLease" [ 1005.670593] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1005.671038] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d10635c0-efb9-43ca-9fe2-1ffc89b6a854 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "fddba36d-1b15-43fb-9e99-68880e8a235c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.633s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.681420] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1005.681420] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52170662-3e1b-030f-0338-0bd11bbac05a" [ 1005.681420] env[61911]: _type = "HttpNfcLease" [ 1005.681420] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1005.736378] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251531, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080856} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.736655] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.737446] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8195cea2-4c37-47cd-9f22-a20dc373e26e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.761716] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] a03413fe-e74d-42e4-83ea-a9c19318526e/a03413fe-e74d-42e4-83ea-a9c19318526e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.762443] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6843cefa-824c-4421-b8f7-9caccec3ab86 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.786620] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1005.786620] env[61911]: value = "task-1251534" [ 1005.786620] env[61911]: _type = "Task" [ 1005.786620] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.796202] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251534, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.921175] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f47b30-53a1-4782-8078-bba947c17987 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.929764] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a5228c-a5ca-4e12-a29d-b4292b84a5fb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.962530] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63becee-b0ba-4113-bdbb-3bfb29a75eb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.970318] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7e2ce2-a075-4332-9985-f788962a4dbd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.984428] env[61911]: DEBUG nova.compute.provider_tree [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.000917] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.090162] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251532, 'name': ReconfigVM_Task, 'duration_secs': 0.488675} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.090480] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269659', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'name': 'volume-19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f63fa31b-8840-4f95-9514-51f7d5393b25', 'attached_at': '', 'detached_at': '', 'volume_id': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9', 'serial': '19d8968b-4a55-4aff-b72f-65e2be6ff7e9'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1006.090786] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.091575] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287f9da4-acd9-4ff8-a8d3-4123e74c60e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.098012] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1006.098250] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70062050-9172-49e7-941f-1cb76f24cbf7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.166776] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1006.167016] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1006.167349] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleting the datastore file [datastore1] f63fa31b-8840-4f95-9514-51f7d5393b25 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1006.167741] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96414ecd-b4e7-4b68-ae29-85fcbb56895a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.176822] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1006.176822] env[61911]: value = "task-1251536" [ 1006.176822] env[61911]: _type = "Task" [ 1006.176822] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.181697] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.181697] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52170662-3e1b-030f-0338-0bd11bbac05a" [ 1006.181697] env[61911]: _type = "HttpNfcLease" [ 1006.181697] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1006.182325] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1006.182325] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52170662-3e1b-030f-0338-0bd11bbac05a" [ 1006.182325] env[61911]: _type = "HttpNfcLease" [ 1006.182325] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1006.183060] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc19feab-86c9-483a-a472-909cbb49e479 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.188521] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251536, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.193162] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5d7df-0da7-1812-3207-fa01319f09de/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1006.193351] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5d7df-0da7-1812-3207-fa01319f09de/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1006.297669] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251534, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.335381] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0a649913-da67-4087-8ee9-0d38bcdf18d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.416320] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "5b7d09e1-ee14-4001-8d73-14e763402670" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.416581] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.487396] env[61911]: DEBUG nova.scheduler.client.report [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1006.688096] env[61911]: DEBUG oslo_vmware.api [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251536, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155344} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.688535] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.688784] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1006.689052] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1006.689369] env[61911]: INFO nova.compute.manager [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Took 2.28 seconds to destroy the instance on the hypervisor. [ 1006.689800] env[61911]: DEBUG oslo.service.loopingcall [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.690112] env[61911]: DEBUG nova.compute.manager [-] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1006.690247] env[61911]: DEBUG nova.network.neutron [-] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1006.797150] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251534, 'name': ReconfigVM_Task, 'duration_secs': 0.817803} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.797472] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Reconfigured VM instance instance-00000061 to attach disk [datastore1] a03413fe-e74d-42e4-83ea-a9c19318526e/a03413fe-e74d-42e4-83ea-a9c19318526e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.798219] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b9292d3-503b-4a73-a469-77e84859c088 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.805380] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1006.805380] env[61911]: value = "task-1251537" [ 1006.805380] env[61911]: _type = "Task" [ 1006.805380] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.816186] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251537, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.920662] env[61911]: INFO nova.compute.manager [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Detaching volume d4fa1f27-0c16-4b2d-9915-9fb3f46332b3 [ 1006.963502] env[61911]: INFO nova.virt.block_device [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Attempting to driver detach volume d4fa1f27-0c16-4b2d-9915-9fb3f46332b3 from mountpoint /dev/sdb [ 1006.963609] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1006.964203] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269629', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'name': 'volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b7d09e1-ee14-4001-8d73-14e763402670', 'attached_at': '', 'detached_at': '', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'serial': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1006.965162] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45989d1e-40de-4d2c-91ee-1c678275043d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.992780] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.859s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.995631] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f737a6a9-954f-4674-a4df-671dab3b106c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.999879] env[61911]: DEBUG nova.compute.manager [req-8bb7d1c0-de39-41e4-8bd3-59a05590af72 req-db1670f6-4917-40a7-a06f-90e214293084 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Received event network-vif-deleted-7438c9cb-caf9-4f7d-b9b9-b9cba545a574 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1007.000310] env[61911]: INFO nova.compute.manager [req-8bb7d1c0-de39-41e4-8bd3-59a05590af72 req-db1670f6-4917-40a7-a06f-90e214293084 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Neutron deleted interface 7438c9cb-caf9-4f7d-b9b9-b9cba545a574; detaching it from the instance and deleting it from the info cache [ 1007.000566] env[61911]: DEBUG nova.network.neutron [req-8bb7d1c0-de39-41e4-8bd3-59a05590af72 req-db1670f6-4917-40a7-a06f-90e214293084 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.002376] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.002s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.003363] env[61911]: DEBUG nova.objects.instance [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lazy-loading 'resources' on Instance uuid f44c597d-ffb8-446b-b814-65733d077dbf {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.013091] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9633c3c2-c324-47a2-9c79-761e37b6dcc1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.037763] env[61911]: INFO nova.scheduler.client.report [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleted allocations for instance b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8 [ 1007.039696] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8bdbf6-2ffd-4227-8df2-c248a9e5b8ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.058438] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] The volume has not been displaced from its original location: [datastore1] volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3/volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1007.064213] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Reconfiguring VM instance instance-0000004b to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1007.065049] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e3fe14d-9f27-47a1-b3eb-9dcf2167a7ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.083523] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1007.083523] env[61911]: value = "task-1251538" [ 1007.083523] env[61911]: _type = "Task" [ 1007.083523] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.093233] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251538, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.318120] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251537, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.456026] env[61911]: DEBUG nova.network.neutron [-] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.509028] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3102e686-4c75-4e06-9111-3b99df00fd58 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.521413] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649eb88a-e0b3-41a8-a43d-5e64b522d2d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.559011] env[61911]: DEBUG nova.compute.manager [req-8bb7d1c0-de39-41e4-8bd3-59a05590af72 req-db1670f6-4917-40a7-a06f-90e214293084 service nova] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Detach interface failed, port_id=7438c9cb-caf9-4f7d-b9b9-b9cba545a574, reason: Instance f63fa31b-8840-4f95-9514-51f7d5393b25 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1007.560103] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8e8a69ce-3c06-49f4-9b56-b40e4b1622a9 tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.690s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.596982] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.671672] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2624596-be7c-4a27-903d-f6ffa318cc4d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.679945] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8668f123-b020-4d5a-b7bd-a5cbf8348a9f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.712196] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4c9f42-d9bb-47f2-9be9-8fc876f10457 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.720400] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e47d93-5e6a-488a-a826-8b1df1205980 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.734238] env[61911]: DEBUG nova.compute.provider_tree [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.817529] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251537, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.880041] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.880419] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.880705] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "b1d2b205-81d7-410b-901d-ccd608bc5bec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.880975] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.881213] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.883600] env[61911]: INFO nova.compute.manager [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Terminating instance [ 1007.958693] env[61911]: INFO nova.compute.manager [-] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Took 1.27 seconds to deallocate network for instance. [ 1008.096560] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.237783] env[61911]: DEBUG nova.scheduler.client.report [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1008.318269] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251537, 'name': Rename_Task, 'duration_secs': 1.176078} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.318858] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1008.319169] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60b1a12b-fc39-48d1-aa67-835443047304 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.325310] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1008.325310] env[61911]: value = "task-1251539" [ 1008.325310] env[61911]: _type = "Task" [ 1008.325310] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.333367] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251539, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.388053] env[61911]: DEBUG nova.compute.manager [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1008.388351] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.389279] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75679fc-1c83-4a0c-94da-d0720929d802 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.397155] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.397430] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9871bc1-9276-46c5-8bf5-a311127b7049 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.403891] env[61911]: DEBUG oslo_vmware.api [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 1008.403891] env[61911]: value = "task-1251540" [ 1008.403891] env[61911]: _type = "Task" [ 1008.403891] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.412391] env[61911]: DEBUG oslo_vmware.api [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251540, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.504359] env[61911]: INFO nova.compute.manager [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Took 0.54 seconds to detach 1 volumes for instance. [ 1008.597185] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251538, 'name': ReconfigVM_Task, 'duration_secs': 1.109453} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.597493] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Reconfigured VM instance instance-0000004b to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1008.602478] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbf46c8b-d3fb-4be8-9559-a27506d86cfe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.617545] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1008.617545] env[61911]: value = "task-1251541" [ 1008.617545] env[61911]: _type = "Task" [ 1008.617545] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.626114] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.743145] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.767995] env[61911]: INFO nova.scheduler.client.report [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Deleted allocations for instance f44c597d-ffb8-446b-b814-65733d077dbf [ 1008.837733] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251539, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.915055] env[61911]: DEBUG oslo_vmware.api [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251540, 'name': PowerOffVM_Task, 'duration_secs': 0.192899} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.915411] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.915586] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.915844] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de917456-5045-46a6-a612-1ce149b50b64 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.979986] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.980245] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.980423] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleting the datastore file [datastore1] b1d2b205-81d7-410b-901d-ccd608bc5bec {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.980729] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01917a85-bf87-4ef0-bd5b-3c7e22efefc9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.988810] env[61911]: DEBUG oslo_vmware.api [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for the task: (returnval){ [ 1008.988810] env[61911]: value = "task-1251543" [ 1008.988810] env[61911]: _type = "Task" [ 1008.988810] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.998076] env[61911]: DEBUG oslo_vmware.api [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251543, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.012372] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.012684] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.012932] env[61911]: DEBUG nova.objects.instance [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'resources' on Instance uuid f63fa31b-8840-4f95-9514-51f7d5393b25 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.128299] env[61911]: DEBUG oslo_vmware.api [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251541, 'name': ReconfigVM_Task, 'duration_secs': 0.136973} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.128638] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269629', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'name': 'volume-d4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b7d09e1-ee14-4001-8d73-14e763402670', 'attached_at': '', 'detached_at': '', 'volume_id': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3', 'serial': 'd4fa1f27-0c16-4b2d-9915-9fb3f46332b3'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1009.276557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ead0d2b6-f8ce-476d-824d-34fe16d0f2bb tempest-ServerDiskConfigTestJSON-921383436 tempest-ServerDiskConfigTestJSON-921383436-project-member] Lock "f44c597d-ffb8-446b-b814-65733d077dbf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.714s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.335294] env[61911]: DEBUG oslo_vmware.api [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251539, 'name': PowerOnVM_Task, 'duration_secs': 0.535435} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.335593] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1009.335807] env[61911]: INFO nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Took 10.48 seconds to spawn the instance on the hypervisor. [ 1009.336014] env[61911]: DEBUG nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1009.336808] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b472bb-66fb-49f9-8072-e9b36d335b46 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.499642] env[61911]: DEBUG oslo_vmware.api [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Task: {'id': task-1251543, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156338} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.499951] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.500179] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.500399] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.500605] env[61911]: INFO nova.compute.manager [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1009.500889] env[61911]: DEBUG oslo.service.loopingcall [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.501138] env[61911]: DEBUG nova.compute.manager [-] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1009.501248] env[61911]: DEBUG nova.network.neutron [-] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.529299] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "c95707ca-7e32-42c0-914a-f0b178f07997" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.529615] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "c95707ca-7e32-42c0-914a-f0b178f07997" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.529878] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "c95707ca-7e32-42c0-914a-f0b178f07997-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.530107] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "c95707ca-7e32-42c0-914a-f0b178f07997-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.530314] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "c95707ca-7e32-42c0-914a-f0b178f07997-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.535088] env[61911]: INFO nova.compute.manager [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Terminating instance [ 1009.644040] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e275d9-c7d7-4573-a5c0-5b097c9cb665 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.651969] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf4a65e-f624-4282-b703-90715268d9a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.687317] env[61911]: DEBUG nova.objects.instance [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'flavor' on Instance uuid 5b7d09e1-ee14-4001-8d73-14e763402670 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.691589] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe39e92f-d179-413a-88aa-2fee8998ef1c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.698306] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf97939-a59d-4d5d-a78d-2dace2293a9a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.714405] env[61911]: DEBUG nova.compute.provider_tree [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.784385] env[61911]: DEBUG nova.compute.manager [req-23c98f9b-5c0b-4c7d-9b24-58202b46155f req-446f0c98-250e-4c12-b605-f95a0aaa6640 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Received event network-vif-deleted-450ba0e3-64ce-4aff-bf9d-059d8aa237fd {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1009.784645] env[61911]: INFO nova.compute.manager [req-23c98f9b-5c0b-4c7d-9b24-58202b46155f req-446f0c98-250e-4c12-b605-f95a0aaa6640 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Neutron deleted interface 450ba0e3-64ce-4aff-bf9d-059d8aa237fd; detaching it from the instance and deleting it from the info cache [ 1009.784847] env[61911]: DEBUG nova.network.neutron [req-23c98f9b-5c0b-4c7d-9b24-58202b46155f req-446f0c98-250e-4c12-b605-f95a0aaa6640 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.856562] env[61911]: INFO nova.compute.manager [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Took 15.56 seconds to build instance. [ 1010.039377] env[61911]: DEBUG nova.compute.manager [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1010.039839] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.040972] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db264173-8b58-4d84-9cbc-1c74b9ed950b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.049248] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.049518] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04defa06-be39-4188-bd27-572070b78a71 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.056946] env[61911]: DEBUG oslo_vmware.api [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 1010.056946] env[61911]: value = "task-1251544" [ 1010.056946] env[61911]: _type = "Task" [ 1010.056946] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.073723] env[61911]: DEBUG oslo_vmware.api [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.219152] env[61911]: DEBUG nova.scheduler.client.report [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1010.262035] env[61911]: DEBUG nova.network.neutron [-] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.287559] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b35302e8-7979-4908-bf7b-4040f649a450 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.297900] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d4ba31-d340-4fdd-bf01-eac508a725af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.333884] env[61911]: DEBUG nova.compute.manager [req-23c98f9b-5c0b-4c7d-9b24-58202b46155f req-446f0c98-250e-4c12-b605-f95a0aaa6640 service nova] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Detach interface failed, port_id=450ba0e3-64ce-4aff-bf9d-059d8aa237fd, reason: Instance b1d2b205-81d7-410b-901d-ccd608bc5bec could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1010.359021] env[61911]: DEBUG oslo_concurrency.lockutils [None req-734fb3ca-864c-44a8-b199-df598ace6fd1 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.071s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.568024] env[61911]: DEBUG oslo_vmware.api [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251544, 'name': PowerOffVM_Task, 'duration_secs': 0.27894} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.569509] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.569509] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.569509] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97d15af1-7ab8-49f2-b164-d7f6f0ebee61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.689826] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.690104] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.690302] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleting the datastore file [datastore2] c95707ca-7e32-42c0-914a-f0b178f07997 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.690647] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49cd7c20-8a5c-4231-a749-3d70e3ac6fa3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.696940] env[61911]: DEBUG oslo_vmware.api [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 1010.696940] env[61911]: value = "task-1251546" [ 1010.696940] env[61911]: _type = "Task" [ 1010.696940] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.700823] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1f3bf4c9-7b33-40a1-b26e-ce7c5aa17486 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.284s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.707174] env[61911]: DEBUG oslo_vmware.api [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251546, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.724089] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.749050] env[61911]: INFO nova.scheduler.client.report [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted allocations for instance f63fa31b-8840-4f95-9514-51f7d5393b25 [ 1010.765252] env[61911]: INFO nova.compute.manager [-] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Took 1.26 seconds to deallocate network for instance. [ 1010.816975] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "a03413fe-e74d-42e4-83ea-a9c19318526e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.817587] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.817728] env[61911]: DEBUG nova.compute.manager [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1010.819081] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669cd919-a356-4862-9a23-e542b80668e2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.827626] env[61911]: DEBUG nova.compute.manager [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61911) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1010.828380] env[61911]: DEBUG nova.objects.instance [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lazy-loading 'flavor' on Instance uuid a03413fe-e74d-42e4-83ea-a9c19318526e {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.216122] env[61911]: DEBUG oslo_vmware.api [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251546, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186768} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.216529] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.216829] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.217566] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.217566] env[61911]: INFO nova.compute.manager [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1011.217908] env[61911]: DEBUG oslo.service.loopingcall [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.218212] env[61911]: DEBUG nova.compute.manager [-] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1011.218362] env[61911]: DEBUG nova.network.neutron [-] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.269644] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9b9ab037-a4ce-4e06-8412-dd287358da72 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "f63fa31b-8840-4f95-9514-51f7d5393b25" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.364s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.271561] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.271878] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.273534] env[61911]: DEBUG nova.objects.instance [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lazy-loading 'resources' on Instance uuid b1d2b205-81d7-410b-901d-ccd608bc5bec {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.766562] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "5b7d09e1-ee14-4001-8d73-14e763402670" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.766887] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.767136] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "5b7d09e1-ee14-4001-8d73-14e763402670-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.767334] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.767536] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.772942] env[61911]: INFO nova.compute.manager [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Terminating instance [ 1011.823335] env[61911]: DEBUG nova.compute.manager [req-af7aaa99-212b-41b9-b768-79b10d9c591e req-e9c485ba-40df-43fd-bf98-ac778fa15372 service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Received event network-vif-deleted-122dbb65-8d1c-4797-ba2f-4c534edd83a8 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1011.823732] env[61911]: INFO nova.compute.manager [req-af7aaa99-212b-41b9-b768-79b10d9c591e req-e9c485ba-40df-43fd-bf98-ac778fa15372 service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Neutron deleted interface 122dbb65-8d1c-4797-ba2f-4c534edd83a8; detaching it from the instance and deleting it from the info cache [ 1011.824102] env[61911]: DEBUG nova.network.neutron [req-af7aaa99-212b-41b9-b768-79b10d9c591e req-e9c485ba-40df-43fd-bf98-ac778fa15372 service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.838880] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.841237] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-985b5542-e4be-4540-ba33-27077383b73e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.853331] env[61911]: DEBUG oslo_vmware.api [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1011.853331] env[61911]: value = "task-1251547" [ 1011.853331] env[61911]: _type = "Task" [ 1011.853331] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.865019] env[61911]: DEBUG oslo_vmware.api [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.956505] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c44142-f50a-48c3-9170-72a8d93783a7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.966229] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4d9d1b-e645-4828-a41d-b215da0e0093 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.998873] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1242793c-99f3-420e-a865-767af51545f2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.008254] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af135fd-bac5-4c64-9819-43b7c7bd0783 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.022116] env[61911]: DEBUG nova.compute.provider_tree [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.185697] env[61911]: DEBUG nova.network.neutron [-] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.279329] env[61911]: DEBUG nova.compute.manager [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1012.279631] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.281293] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead06edf-610e-4728-9059-be263fbfeae7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.289801] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.290094] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33816446-b2c5-43d9-a9c1-21f5dde6fff2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.296677] env[61911]: DEBUG oslo_vmware.api [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1012.296677] env[61911]: value = "task-1251548" [ 1012.296677] env[61911]: _type = "Task" [ 1012.296677] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.306969] env[61911]: DEBUG oslo_vmware.api [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.330093] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-156d7a20-4203-42dd-aed8-3822eebdd700 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.339070] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d52552-24a3-4475-a031-2eef9dee38a5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.371621] env[61911]: DEBUG nova.compute.manager [req-af7aaa99-212b-41b9-b768-79b10d9c591e req-e9c485ba-40df-43fd-bf98-ac778fa15372 service nova] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Detach interface failed, port_id=122dbb65-8d1c-4797-ba2f-4c534edd83a8, reason: Instance c95707ca-7e32-42c0-914a-f0b178f07997 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1012.379953] env[61911]: DEBUG oslo_vmware.api [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251547, 'name': PowerOffVM_Task, 'duration_secs': 0.242926} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.380288] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.380531] env[61911]: DEBUG nova.compute.manager [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1012.381421] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362251d5-460e-43ea-a0d1-4b551c2058fd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.526178] env[61911]: DEBUG nova.scheduler.client.report [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.689925] env[61911]: INFO nova.compute.manager [-] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Took 1.47 seconds to deallocate network for instance. [ 1012.807839] env[61911]: DEBUG oslo_vmware.api [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251548, 'name': PowerOffVM_Task, 'duration_secs': 0.190773} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.807839] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.807839] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1012.808438] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09502eaa-0b5b-4d55-ae01-b9ae7368cdc3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.873040] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.873040] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.873388] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleting the datastore file [datastore1] 5b7d09e1-ee14-4001-8d73-14e763402670 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.873723] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c036c21-4b63-4d1e-9257-2542a25fedcf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.880307] env[61911]: DEBUG oslo_vmware.api [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1012.880307] env[61911]: value = "task-1251550" [ 1012.880307] env[61911]: _type = "Task" [ 1012.880307] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.888878] env[61911]: DEBUG oslo_vmware.api [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.893824] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7b88b0d4-b2a5-4541-a04c-6316889a0df8 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.076s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.002399] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5d7df-0da7-1812-3207-fa01319f09de/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1013.003428] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2362a6-b714-48f0-8e1b-b4e4c3216af6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.010299] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5d7df-0da7-1812-3207-fa01319f09de/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1013.010524] env[61911]: ERROR oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5d7df-0da7-1812-3207-fa01319f09de/disk-0.vmdk due to incomplete transfer. [ 1013.010665] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ce8e9007-efdd-4974-b31d-706dc21fca10 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.017821] env[61911]: DEBUG oslo_vmware.rw_handles [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5d7df-0da7-1812-3207-fa01319f09de/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1013.018047] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Uploaded image 8cdf5b0c-7e3c-4b1b-8bfc-345854bccdff to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1013.020358] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1013.020593] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-55a110b8-0eac-44df-8feb-98ccfbaa8fa0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.027034] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1013.027034] env[61911]: value = "task-1251551" [ 1013.027034] env[61911]: _type = "Task" [ 1013.027034] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.031824] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.037754] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251551, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.054907] env[61911]: INFO nova.scheduler.client.report [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Deleted allocations for instance b1d2b205-81d7-410b-901d-ccd608bc5bec [ 1013.199666] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.200026] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.200323] env[61911]: DEBUG nova.objects.instance [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lazy-loading 'resources' on Instance uuid c95707ca-7e32-42c0-914a-f0b178f07997 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.201397] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.201612] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.392673] env[61911]: DEBUG oslo_vmware.api [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136946} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.392952] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.393203] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1013.393435] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1013.394698] env[61911]: INFO nova.compute.manager [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1013.394698] env[61911]: DEBUG oslo.service.loopingcall [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.394698] env[61911]: DEBUG nova.compute.manager [-] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1013.394698] env[61911]: DEBUG nova.network.neutron [-] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1013.537168] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251551, 'name': Destroy_Task, 'duration_secs': 0.313566} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.537473] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Destroyed the VM [ 1013.538368] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1013.538368] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-59203a31-1d71-458d-b4fe-127b25835a8a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.543653] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1013.543653] env[61911]: value = "task-1251552" [ 1013.543653] env[61911]: _type = "Task" [ 1013.543653] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.551485] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251552, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.564892] env[61911]: DEBUG oslo_concurrency.lockutils [None req-fa9f28b2-df9a-4b95-af5a-0df2441f3e9b tempest-ListServerFiltersTestJSON-95480912 tempest-ListServerFiltersTestJSON-95480912-project-member] Lock "b1d2b205-81d7-410b-901d-ccd608bc5bec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.684s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.611683] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "a03413fe-e74d-42e4-83ea-a9c19318526e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.611973] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.612208] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "a03413fe-e74d-42e4-83ea-a9c19318526e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.612407] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.612694] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.614942] env[61911]: INFO nova.compute.manager [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Terminating instance [ 1013.708187] env[61911]: DEBUG nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1013.868088] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724c9ba2-2ea0-4c6e-90f6-7ee143b99f59 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.876126] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1e680a-b95d-4d42-a179-109de63a73be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.913373] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bd01f9-c1e3-45ef-a523-942c35fee223 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.922989] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d265590-20ac-40e6-956c-84d6583d5aa9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.933044] env[61911]: DEBUG nova.compute.manager [req-0805b679-55a0-4d42-98aa-979a80046526 req-1c81931b-0348-4fa7-8d2a-c81346e59da4 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Received event network-vif-deleted-5c7df9be-477d-4aab-9d5b-e340866dc6bb {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1013.933044] env[61911]: INFO nova.compute.manager [req-0805b679-55a0-4d42-98aa-979a80046526 req-1c81931b-0348-4fa7-8d2a-c81346e59da4 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Neutron deleted interface 5c7df9be-477d-4aab-9d5b-e340866dc6bb; detaching it from the instance and deleting it from the info cache [ 1013.933044] env[61911]: DEBUG nova.network.neutron [req-0805b679-55a0-4d42-98aa-979a80046526 req-1c81931b-0348-4fa7-8d2a-c81346e59da4 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.941296] env[61911]: DEBUG nova.compute.provider_tree [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.053728] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251552, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.120258] env[61911]: DEBUG nova.compute.manager [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1014.120423] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.121501] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8188f0d-e7da-4c66-873e-bc8d5defa7f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.129219] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.129492] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7742e9e7-48f7-4a4c-a147-1aac68a2c5e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.228227] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.417618] env[61911]: DEBUG nova.network.neutron [-] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.432469] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e529eb9-f05b-4485-b2e8-e8a25307c56a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.446259] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9e0fba-efe0-464e-b9c6-fceb2e52c558 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.482194] env[61911]: DEBUG nova.compute.manager [req-0805b679-55a0-4d42-98aa-979a80046526 req-1c81931b-0348-4fa7-8d2a-c81346e59da4 service nova] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Detach interface failed, port_id=5c7df9be-477d-4aab-9d5b-e340866dc6bb, reason: Instance 5b7d09e1-ee14-4001-8d73-14e763402670 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1014.483498] env[61911]: ERROR nova.scheduler.client.report [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [req-160a1be6-dc35-483f-be19-45662cd02fdb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID b8a56394-897a-4e67-92ba-db46db6115e9. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-160a1be6-dc35-483f-be19-45662cd02fdb"}]} [ 1014.504704] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.504943] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.505153] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleting the datastore file [datastore1] a03413fe-e74d-42e4-83ea-a9c19318526e {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.506488] env[61911]: DEBUG nova.scheduler.client.report [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Refreshing inventories for resource provider b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1014.508378] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc348b0e-c485-46ad-a49c-7afd701ebe4c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.515701] env[61911]: DEBUG oslo_vmware.api [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1014.515701] env[61911]: value = "task-1251554" [ 1014.515701] env[61911]: _type = "Task" [ 1014.515701] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.526487] env[61911]: DEBUG oslo_vmware.api [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251554, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.531233] env[61911]: DEBUG nova.scheduler.client.report [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Updating ProviderTree inventory for provider b8a56394-897a-4e67-92ba-db46db6115e9 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1014.531516] env[61911]: DEBUG nova.compute.provider_tree [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.550271] env[61911]: DEBUG nova.scheduler.client.report [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Refreshing aggregate associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, aggregates: None {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1014.557485] env[61911]: DEBUG oslo_vmware.api [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251552, 'name': RemoveSnapshot_Task, 'duration_secs': 0.858231} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.558138] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1014.558138] env[61911]: INFO nova.compute.manager [None req-1368c466-7dff-4a2c-84c5-b6094afa8d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Took 12.55 seconds to snapshot the instance on the hypervisor. [ 1014.572860] env[61911]: DEBUG nova.scheduler.client.report [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Refreshing trait associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1014.709250] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39450230-79fc-43d9-afa5-8d4cea8498e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.717556] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbae80de-79f6-467a-88a9-da5ea0edb05e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.754478] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a50bd99-e542-4fc3-ab05-1c4c8f20ead9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.762615] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4a9a04-f403-459d-9b5d-4f4a1651c294 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.777152] env[61911]: DEBUG nova.compute.provider_tree [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.926140] env[61911]: INFO nova.compute.manager [-] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Took 1.53 seconds to deallocate network for instance. [ 1015.027015] env[61911]: DEBUG oslo_vmware.api [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251554, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141406} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.027314] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.027511] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1015.027764] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1015.027873] env[61911]: INFO nova.compute.manager [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Took 0.91 seconds to destroy the instance on the hypervisor. [ 1015.028139] env[61911]: DEBUG oslo.service.loopingcall [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.028346] env[61911]: DEBUG nova.compute.manager [-] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1015.028441] env[61911]: DEBUG nova.network.neutron [-] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1015.319766] env[61911]: DEBUG nova.scheduler.client.report [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Updated inventory for provider b8a56394-897a-4e67-92ba-db46db6115e9 with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1015.320188] env[61911]: DEBUG nova.compute.provider_tree [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Updating resource provider b8a56394-897a-4e67-92ba-db46db6115e9 generation from 115 to 116 during operation: update_inventory {{(pid=61911) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1015.323091] env[61911]: DEBUG nova.compute.provider_tree [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1015.434880] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.819522] env[61911]: DEBUG nova.network.neutron [-] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.826698] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.625s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.828848] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.600s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.830077] env[61911]: INFO nova.compute.claims [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.858829] env[61911]: INFO nova.scheduler.client.report [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleted allocations for instance c95707ca-7e32-42c0-914a-f0b178f07997 [ 1015.958407] env[61911]: DEBUG nova.compute.manager [req-f1f16cc4-13d3-48fc-8ddd-c9da26207f66 req-12ed11f1-52e5-4451-864a-1a55a88fa475 service nova] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Received event network-vif-deleted-73e4548b-9760-484e-9023-f467279085d6 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1016.259650] env[61911]: DEBUG nova.compute.manager [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1016.259859] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16da8e0-d954-48d0-8e67-933ae2951060 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.319164] env[61911]: INFO nova.compute.manager [-] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Took 1.29 seconds to deallocate network for instance. [ 1016.364883] env[61911]: DEBUG oslo_concurrency.lockutils [None req-64718a9b-046a-427f-8818-2f3c7b47af7e tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "c95707ca-7e32-42c0-914a-f0b178f07997" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.835s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.772152] env[61911]: INFO nova.compute.manager [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] instance snapshotting [ 1016.778033] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4609b3cb-f137-4304-b150-10c04d3d1f51 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.794278] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0146936-97ad-4915-8197-c525880040ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.830181] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.958033] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bbea81-99da-4b94-be2e-4907220a2109 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.977602] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909773a7-e584-4532-bab8-c46cb1346807 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.054870] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850c395b-d608-47f4-941e-5bcc0216d3c2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.072861] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3d4f55-24c5-4425-be22-112988dcafcb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.105224] env[61911]: DEBUG nova.compute.provider_tree [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.228797] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.228952] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.306045] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1017.306045] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7e16f887-ed35-4a2b-a82b-9212f66ee36f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.313659] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1017.313659] env[61911]: value = "task-1251555" [ 1017.313659] env[61911]: _type = "Task" [ 1017.313659] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.322322] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251555, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.607779] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "bafd58bc-8ae7-49a4-a039-31d328699010" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.608098] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "bafd58bc-8ae7-49a4-a039-31d328699010" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.608341] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "bafd58bc-8ae7-49a4-a039-31d328699010-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.608542] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "bafd58bc-8ae7-49a4-a039-31d328699010-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.608723] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "bafd58bc-8ae7-49a4-a039-31d328699010-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.611107] env[61911]: DEBUG nova.scheduler.client.report [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1017.614334] env[61911]: INFO nova.compute.manager [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Terminating instance [ 1017.731594] env[61911]: DEBUG nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1017.823401] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251555, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.119976] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.120570] env[61911]: DEBUG nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1018.123264] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.688s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.123364] env[61911]: DEBUG nova.objects.instance [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'resources' on Instance uuid 5b7d09e1-ee14-4001-8d73-14e763402670 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.124739] env[61911]: DEBUG nova.compute.manager [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1018.124835] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.125963] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ea28b8-d9af-43f5-ab83-2c8ae8065322 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.134057] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.134180] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fddefe8-97ce-47ef-ab9a-191adc7be011 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.141168] env[61911]: DEBUG oslo_vmware.api [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 1018.141168] env[61911]: value = "task-1251556" [ 1018.141168] env[61911]: _type = "Task" [ 1018.141168] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.149154] env[61911]: DEBUG oslo_vmware.api [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.254263] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.324978] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251555, 'name': CreateSnapshot_Task, 'duration_secs': 0.633968} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.325358] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1018.326262] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a58e689-0f79-418a-8723-9475f51315b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.629957] env[61911]: DEBUG nova.compute.utils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.631445] env[61911]: DEBUG nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1018.631612] env[61911]: DEBUG nova.network.neutron [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1018.652074] env[61911]: DEBUG oslo_vmware.api [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251556, 'name': PowerOffVM_Task, 'duration_secs': 0.316392} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.652347] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.652516] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.652763] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-899b16ef-d871-4468-9f9f-edeee271818e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.681374] env[61911]: DEBUG nova.policy [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a5252f9309f4204a3a3fd7d865db6e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b285793da304b5e8f9736098ef163c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1018.714102] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.714337] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.714527] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleting the datastore file [datastore2] bafd58bc-8ae7-49a4-a039-31d328699010 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.714827] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccb01456-efdd-4cd5-b506-7a2b11180c81 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.721314] env[61911]: DEBUG oslo_vmware.api [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for the task: (returnval){ [ 1018.721314] env[61911]: value = "task-1251558" [ 1018.721314] env[61911]: _type = "Task" [ 1018.721314] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.731212] env[61911]: DEBUG oslo_vmware.api [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251558, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.734915] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c29f84-effa-4e34-bf2b-0ae8772eb729 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.741517] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7714a823-497b-4252-abaa-d1601b4af8db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.773290] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef45dbe3-a7ec-4cef-9875-6dcbd83f6b0f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.781507] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96493f71-e519-41e3-863c-b1d76bc5fd7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.795785] env[61911]: DEBUG nova.compute.provider_tree [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.844424] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1018.844697] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d9a2222f-63d7-460f-9785-2faaf77149c2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.853519] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1018.853519] env[61911]: value = "task-1251559" [ 1018.853519] env[61911]: _type = "Task" [ 1018.853519] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.860958] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251559, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.923086] env[61911]: DEBUG nova.network.neutron [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Successfully created port: 92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.134250] env[61911]: DEBUG nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1019.234576] env[61911]: DEBUG oslo_vmware.api [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Task: {'id': task-1251558, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17352} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.234901] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.235127] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.235301] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.235494] env[61911]: INFO nova.compute.manager [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1019.235778] env[61911]: DEBUG oslo.service.loopingcall [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.235966] env[61911]: DEBUG nova.compute.manager [-] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1019.236024] env[61911]: DEBUG nova.network.neutron [-] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1019.299067] env[61911]: DEBUG nova.scheduler.client.report [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1019.367093] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251559, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.522951] env[61911]: DEBUG nova.compute.manager [req-0c41b9b4-9759-4976-adf5-86853c0c3ca5 req-431d6940-5518-48a4-bd67-86c21761425d service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Received event network-vif-deleted-15e642fd-30d2-41c2-98f7-0d19e729041e {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1019.522951] env[61911]: INFO nova.compute.manager [req-0c41b9b4-9759-4976-adf5-86853c0c3ca5 req-431d6940-5518-48a4-bd67-86c21761425d service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Neutron deleted interface 15e642fd-30d2-41c2-98f7-0d19e729041e; detaching it from the instance and deleting it from the info cache [ 1019.523120] env[61911]: DEBUG nova.network.neutron [req-0c41b9b4-9759-4976-adf5-86853c0c3ca5 req-431d6940-5518-48a4-bd67-86c21761425d service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.805739] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.808311] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.978s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.808572] env[61911]: DEBUG nova.objects.instance [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lazy-loading 'resources' on Instance uuid a03413fe-e74d-42e4-83ea-a9c19318526e {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.831059] env[61911]: INFO nova.scheduler.client.report [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted allocations for instance 5b7d09e1-ee14-4001-8d73-14e763402670 [ 1019.865780] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251559, 'name': CloneVM_Task} progress is 95%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.998826] env[61911]: DEBUG nova.network.neutron [-] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.026286] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e508b34-5270-4d51-aa7a-41c464a03bf8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.035016] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59779184-1ff8-4777-9725-d4d5f56d0704 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.060597] env[61911]: DEBUG nova.compute.manager [req-0c41b9b4-9759-4976-adf5-86853c0c3ca5 req-431d6940-5518-48a4-bd67-86c21761425d service nova] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Detach interface failed, port_id=15e642fd-30d2-41c2-98f7-0d19e729041e, reason: Instance bafd58bc-8ae7-49a4-a039-31d328699010 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1020.145217] env[61911]: DEBUG nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1020.166160] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1020.166416] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.166582] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1020.166851] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.167089] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1020.167277] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1020.167492] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1020.167661] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1020.168268] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1020.168560] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1020.168778] env[61911]: DEBUG nova.virt.hardware [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1020.169640] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aad3d9f-778f-46d3-938b-d29f8ed06f16 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.177715] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cb81bc-822b-4b2d-8c8b-fd6ec7d35497 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.338807] env[61911]: DEBUG nova.compute.manager [req-d63c1d58-2d21-4851-9859-f342592d435b req-3f984366-785e-4741-afb2-3b802158a351 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Received event network-vif-plugged-92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1020.339117] env[61911]: DEBUG oslo_concurrency.lockutils [req-d63c1d58-2d21-4851-9859-f342592d435b req-3f984366-785e-4741-afb2-3b802158a351 service nova] Acquiring lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.339281] env[61911]: DEBUG oslo_concurrency.lockutils [req-d63c1d58-2d21-4851-9859-f342592d435b req-3f984366-785e-4741-afb2-3b802158a351 service nova] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.339487] env[61911]: DEBUG oslo_concurrency.lockutils [req-d63c1d58-2d21-4851-9859-f342592d435b req-3f984366-785e-4741-afb2-3b802158a351 service nova] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.339625] env[61911]: DEBUG nova.compute.manager [req-d63c1d58-2d21-4851-9859-f342592d435b req-3f984366-785e-4741-afb2-3b802158a351 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] No waiting events found dispatching network-vif-plugged-92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1020.339798] env[61911]: WARNING nova.compute.manager [req-d63c1d58-2d21-4851-9859-f342592d435b req-3f984366-785e-4741-afb2-3b802158a351 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Received unexpected event network-vif-plugged-92b158ae-a674-4ba7-a4a5-c4819bba4492 for instance with vm_state building and task_state spawning. [ 1020.342416] env[61911]: DEBUG oslo_concurrency.lockutils [None req-bc5aa80a-c8a3-470b-a940-2d2065378db1 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "5b7d09e1-ee14-4001-8d73-14e763402670" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.576s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.364739] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251559, 'name': CloneVM_Task, 'duration_secs': 1.218117} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.367210] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Created linked-clone VM from snapshot [ 1020.368523] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bd3ece-cb33-4834-ada8-65a80e8df545 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.379334] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Uploading image dabc0dd7-8ea1-48f4-8766-664d76c32e4a {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1020.412577] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1020.412577] env[61911]: value = "vm-269664" [ 1020.412577] env[61911]: _type = "VirtualMachine" [ 1020.412577] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1020.413194] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f2033399-d5d5-4018-8bb5-9fdb544d8c1b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.422939] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lease: (returnval){ [ 1020.422939] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e30327-e402-e83f-36f9-5a00132c4a54" [ 1020.422939] env[61911]: _type = "HttpNfcLease" [ 1020.422939] env[61911]: } obtained for exporting VM: (result){ [ 1020.422939] env[61911]: value = "vm-269664" [ 1020.422939] env[61911]: _type = "VirtualMachine" [ 1020.422939] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1020.423885] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the lease: (returnval){ [ 1020.423885] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e30327-e402-e83f-36f9-5a00132c4a54" [ 1020.423885] env[61911]: _type = "HttpNfcLease" [ 1020.423885] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1020.431072] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1020.431072] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e30327-e402-e83f-36f9-5a00132c4a54" [ 1020.431072] env[61911]: _type = "HttpNfcLease" [ 1020.431072] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1020.445206] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4383f5d-666a-44b4-bc02-83c7cf548be6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.452423] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2d15b5-b810-4857-b118-941a122c9891 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.484085] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e786bcef-16ef-4090-b734-0dde54edb46e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.491293] env[61911]: DEBUG nova.network.neutron [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Successfully updated port: 92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.493390] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b869e9b-c279-4787-be46-2ae30576c7e8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.506990] env[61911]: INFO nova.compute.manager [-] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Took 1.27 seconds to deallocate network for instance. [ 1020.508336] env[61911]: DEBUG nova.compute.provider_tree [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.933161] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1020.933161] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e30327-e402-e83f-36f9-5a00132c4a54" [ 1020.933161] env[61911]: _type = "HttpNfcLease" [ 1020.933161] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1020.933490] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1020.933490] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e30327-e402-e83f-36f9-5a00132c4a54" [ 1020.933490] env[61911]: _type = "HttpNfcLease" [ 1020.933490] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1020.934419] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30cfeeb-84de-40fc-a7ea-d86b7a3c16fa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.942277] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5260464f-9253-b891-7ad5-561e15be94d3/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1020.942461] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5260464f-9253-b891-7ad5-561e15be94d3/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1021.000174] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.000336] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.000495] env[61911]: DEBUG nova.network.neutron [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.012730] env[61911]: DEBUG nova.scheduler.client.report [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1021.017409] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.039475] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-df56782f-50e0-4602-9d91-fe70b0533df8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.517491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.709s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.521133] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.266s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.523428] env[61911]: INFO nova.compute.claims [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.539546] env[61911]: DEBUG nova.network.neutron [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1021.567833] env[61911]: INFO nova.scheduler.client.report [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted allocations for instance a03413fe-e74d-42e4-83ea-a9c19318526e [ 1021.690729] env[61911]: DEBUG nova.network.neutron [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Updating instance_info_cache with network_info: [{"id": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "address": "fa:16:3e:58:3a:db", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b158ae-a6", "ovs_interfaceid": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.078344] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d903bc54-503b-4d12-9299-bbac63cb48d2 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "a03413fe-e74d-42e4-83ea-a9c19318526e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.466s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.193417] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.193870] env[61911]: DEBUG nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Instance network_info: |[{"id": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "address": "fa:16:3e:58:3a:db", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b158ae-a6", "ovs_interfaceid": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1022.194391] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:3a:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92b158ae-a674-4ba7-a4a5-c4819bba4492', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.203140] env[61911]: DEBUG oslo.service.loopingcall [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.203467] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.204249] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-499df01e-9eec-4d4d-ad40-a5209f42d916 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.224530] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.224530] env[61911]: value = "task-1251561" [ 1022.224530] env[61911]: _type = "Task" [ 1022.224530] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.232959] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251561, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.364912] env[61911]: DEBUG nova.compute.manager [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Received event network-changed-92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1022.365177] env[61911]: DEBUG nova.compute.manager [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Refreshing instance network info cache due to event network-changed-92b158ae-a674-4ba7-a4a5-c4819bba4492. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1022.365422] env[61911]: DEBUG oslo_concurrency.lockutils [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] Acquiring lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.365596] env[61911]: DEBUG oslo_concurrency.lockutils [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] Acquired lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.365771] env[61911]: DEBUG nova.network.neutron [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Refreshing network info cache for port 92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1022.563664] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.563982] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.654120] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d6ee81-ac96-405c-9fc7-8d9b406c1203 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.661800] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa70400-b27e-4624-bfcc-70b978d9233c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.696723] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c284e4-b819-4ee0-aabf-6970ad116866 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.704517] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7123cdce-d88b-4e61-b65b-0aab26869b24 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.718883] env[61911]: DEBUG nova.compute.provider_tree [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.734326] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251561, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.066435] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "31f0a4f7-a69d-4ec4-966d-93f402163b88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.066714] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.066954] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "31f0a4f7-a69d-4ec4-966d-93f402163b88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.067176] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.067362] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.069288] env[61911]: DEBUG nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1023.072432] env[61911]: INFO nova.compute.manager [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Terminating instance [ 1023.077274] env[61911]: DEBUG nova.network.neutron [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Updated VIF entry in instance network info cache for port 92b158ae-a674-4ba7-a4a5-c4819bba4492. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1023.077643] env[61911]: DEBUG nova.network.neutron [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Updating instance_info_cache with network_info: [{"id": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "address": "fa:16:3e:58:3a:db", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b158ae-a6", "ovs_interfaceid": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.224615] env[61911]: DEBUG nova.scheduler.client.report [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1023.237386] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251561, 'name': CreateVM_Task, 'duration_secs': 0.518701} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.237647] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1023.238581] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.238784] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.239235] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1023.240072] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49a2794e-7123-41da-8051-2b48a93ebb42 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.245598] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1023.245598] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52662259-5e15-ee83-e5d6-6e6263165a21" [ 1023.245598] env[61911]: _type = "Task" [ 1023.245598] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.254480] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52662259-5e15-ee83-e5d6-6e6263165a21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.577622] env[61911]: DEBUG nova.compute.manager [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1023.578190] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1023.581094] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0f5f33-6a69-4f9b-a8b9-a28353edaa0c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.584692] env[61911]: DEBUG oslo_concurrency.lockutils [req-9962a4ae-7c71-415b-9e64-dd2391116871 req-b744d7a5-9402-4e13-a9db-a73b27bbd7df service nova] Releasing lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.590008] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1023.590328] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65ae74ee-fe80-4872-93b7-e2603ef53be6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.596834] env[61911]: DEBUG oslo_vmware.api [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1023.596834] env[61911]: value = "task-1251562" [ 1023.596834] env[61911]: _type = "Task" [ 1023.596834] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.597967] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.606647] env[61911]: DEBUG oslo_vmware.api [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.733028] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.213s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.733438] env[61911]: DEBUG nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1023.736447] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.719s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.736553] env[61911]: DEBUG nova.objects.instance [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lazy-loading 'resources' on Instance uuid bafd58bc-8ae7-49a4-a039-31d328699010 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.758273] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52662259-5e15-ee83-e5d6-6e6263165a21, 'name': SearchDatastore_Task, 'duration_secs': 0.009328} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.758693] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.759020] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.759339] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.759545] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.759822] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.760161] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43d04bce-f9a2-4887-9ef9-52d2edff7982 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.769851] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.770070] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.771553] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80102441-0d2e-4b1b-a8bf-e09354e6cd55 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.778309] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1023.778309] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52623c9b-b281-1e78-ded6-3d821bba7c40" [ 1023.778309] env[61911]: _type = "Task" [ 1023.778309] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.788653] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52623c9b-b281-1e78-ded6-3d821bba7c40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.108020] env[61911]: DEBUG oslo_vmware.api [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251562, 'name': PowerOffVM_Task, 'duration_secs': 0.235293} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.108345] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1024.108528] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1024.108787] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e50af6f-003f-4563-b1d1-18ab64861377 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.169057] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1024.169421] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1024.169628] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleting the datastore file [datastore2] 31f0a4f7-a69d-4ec4-966d-93f402163b88 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.169922] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc6346c2-42f6-43b7-aa61-d66fd5767f0c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.176139] env[61911]: DEBUG oslo_vmware.api [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for the task: (returnval){ [ 1024.176139] env[61911]: value = "task-1251564" [ 1024.176139] env[61911]: _type = "Task" [ 1024.176139] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.184801] env[61911]: DEBUG oslo_vmware.api [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251564, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.239797] env[61911]: DEBUG nova.compute.utils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.244443] env[61911]: DEBUG nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1024.244873] env[61911]: DEBUG nova.network.neutron [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1024.292384] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52623c9b-b281-1e78-ded6-3d821bba7c40, 'name': SearchDatastore_Task, 'duration_secs': 0.011161} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.293927] env[61911]: DEBUG nova.policy [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc117e32967d4861a77259c18e0c9c33', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce8675694c2841a58e87bb7250f3435a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.298383] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1608cd94-38d7-479d-b066-1ec9efb4d559 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.304991] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1024.304991] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f98d63-ef00-8d40-50f0-8497028a1378" [ 1024.304991] env[61911]: _type = "Task" [ 1024.304991] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.312775] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f98d63-ef00-8d40-50f0-8497028a1378, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.347747] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d1dd96-be6d-4cde-b006-dd879ea5a624 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.356896] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b2d4c6-2908-48c4-ab59-632b7b557505 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.387569] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844c22d8-bdcc-4bb9-9bc6-7ad77fb60bc0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.395348] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13eaa3df-457c-4eb7-af80-0bc681ed6100 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.410607] env[61911]: DEBUG nova.compute.provider_tree [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.592829] env[61911]: DEBUG nova.network.neutron [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Successfully created port: 19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1024.687183] env[61911]: DEBUG oslo_vmware.api [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Task: {'id': task-1251564, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13844} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.687444] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.687641] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.687846] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.688089] env[61911]: INFO nova.compute.manager [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1024.688401] env[61911]: DEBUG oslo.service.loopingcall [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.688611] env[61911]: DEBUG nova.compute.manager [-] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1024.688702] env[61911]: DEBUG nova.network.neutron [-] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1024.745674] env[61911]: DEBUG nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1024.815931] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f98d63-ef00-8d40-50f0-8497028a1378, 'name': SearchDatastore_Task, 'duration_secs': 0.023593} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.816553] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.816836] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] d4e33bff-8d26-49de-98c8-d9dcbfc640dd/d4e33bff-8d26-49de-98c8-d9dcbfc640dd.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1024.817263] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3f26713-3ffa-4612-9fce-102d00ceb6b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.825348] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1024.825348] env[61911]: value = "task-1251565" [ 1024.825348] env[61911]: _type = "Task" [ 1024.825348] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.835136] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251565, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.914200] env[61911]: DEBUG nova.scheduler.client.report [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1025.028631] env[61911]: DEBUG nova.compute.manager [req-ecff284b-dabc-4863-93e4-0dc29cb8f7e8 req-2bdc97de-607b-4e61-beb9-a0e1523898cf service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Received event network-vif-deleted-8d69ab4c-3b99-412f-8985-3d8a9210f776 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1025.028864] env[61911]: INFO nova.compute.manager [req-ecff284b-dabc-4863-93e4-0dc29cb8f7e8 req-2bdc97de-607b-4e61-beb9-a0e1523898cf service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Neutron deleted interface 8d69ab4c-3b99-412f-8985-3d8a9210f776; detaching it from the instance and deleting it from the info cache [ 1025.029290] env[61911]: DEBUG nova.network.neutron [req-ecff284b-dabc-4863-93e4-0dc29cb8f7e8 req-2bdc97de-607b-4e61-beb9-a0e1523898cf service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.335903] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251565, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4342} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.337045] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] d4e33bff-8d26-49de-98c8-d9dcbfc640dd/d4e33bff-8d26-49de-98c8-d9dcbfc640dd.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1025.337045] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.337045] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a864884-2c99-469e-af8c-cee284d467f2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.342828] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1025.342828] env[61911]: value = "task-1251566" [ 1025.342828] env[61911]: _type = "Task" [ 1025.342828] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.350762] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251566, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.419809] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.422270] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.824s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.423952] env[61911]: INFO nova.compute.claims [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.443930] env[61911]: INFO nova.scheduler.client.report [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Deleted allocations for instance bafd58bc-8ae7-49a4-a039-31d328699010 [ 1025.488571] env[61911]: DEBUG nova.network.neutron [-] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.532498] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9478099-a23f-435d-8973-ae0db1f04155 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.541789] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51f4062-0ae2-4f21-855d-4b5f491953e2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.568510] env[61911]: DEBUG nova.compute.manager [req-ecff284b-dabc-4863-93e4-0dc29cb8f7e8 req-2bdc97de-607b-4e61-beb9-a0e1523898cf service nova] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Detach interface failed, port_id=8d69ab4c-3b99-412f-8985-3d8a9210f776, reason: Instance 31f0a4f7-a69d-4ec4-966d-93f402163b88 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1025.755915] env[61911]: DEBUG nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1025.778551] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1025.778863] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.779051] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1025.779239] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.779396] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1025.779578] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1025.779767] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1025.779931] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1025.780118] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1025.780288] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1025.780500] env[61911]: DEBUG nova.virt.hardware [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1025.781387] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbbc36b-14ef-4276-b0d0-3643402b405a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.789312] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ef1c75-435e-4566-9dd8-e4c5802bfe25 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.852244] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251566, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08048} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.852520] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.853314] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772e73ed-cee4-4fc0-a493-da044f83fbfa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.875238] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d4e33bff-8d26-49de-98c8-d9dcbfc640dd/d4e33bff-8d26-49de-98c8-d9dcbfc640dd.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.875561] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00a4dc0d-e5d7-4eef-b0f9-e7cbdf3b0cfa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.895230] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1025.895230] env[61911]: value = "task-1251567" [ 1025.895230] env[61911]: _type = "Task" [ 1025.895230] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.904022] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.951526] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8a5a028c-0ea6-4dce-8597-5671cf3ab397 tempest-ServerRescueNegativeTestJSON-709602546 tempest-ServerRescueNegativeTestJSON-709602546-project-member] Lock "bafd58bc-8ae7-49a4-a039-31d328699010" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.343s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.990997] env[61911]: INFO nova.compute.manager [-] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Took 1.30 seconds to deallocate network for instance. [ 1026.099165] env[61911]: DEBUG nova.compute.manager [req-0255f333-38eb-4be0-8be7-46b25f54bdc3 req-a4a2dc66-935b-407d-a4a1-5965fa9f190a service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-vif-plugged-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1026.099408] env[61911]: DEBUG oslo_concurrency.lockutils [req-0255f333-38eb-4be0-8be7-46b25f54bdc3 req-a4a2dc66-935b-407d-a4a1-5965fa9f190a service nova] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.099642] env[61911]: DEBUG oslo_concurrency.lockutils [req-0255f333-38eb-4be0-8be7-46b25f54bdc3 req-a4a2dc66-935b-407d-a4a1-5965fa9f190a service nova] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.099828] env[61911]: DEBUG oslo_concurrency.lockutils [req-0255f333-38eb-4be0-8be7-46b25f54bdc3 req-a4a2dc66-935b-407d-a4a1-5965fa9f190a service nova] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.100016] env[61911]: DEBUG nova.compute.manager [req-0255f333-38eb-4be0-8be7-46b25f54bdc3 req-a4a2dc66-935b-407d-a4a1-5965fa9f190a service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] No waiting events found dispatching network-vif-plugged-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1026.100207] env[61911]: WARNING nova.compute.manager [req-0255f333-38eb-4be0-8be7-46b25f54bdc3 req-a4a2dc66-935b-407d-a4a1-5965fa9f190a service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received unexpected event network-vif-plugged-19140385-696d-4e4c-a2dd-d53c936d4416 for instance with vm_state building and task_state spawning. [ 1026.286835] env[61911]: DEBUG nova.network.neutron [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Successfully updated port: 19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.405639] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251567, 'name': ReconfigVM_Task, 'duration_secs': 0.497934} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.405639] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d4e33bff-8d26-49de-98c8-d9dcbfc640dd/d4e33bff-8d26-49de-98c8-d9dcbfc640dd.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.406051] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f22fb8f6-a457-40b6-aad3-5b530644021a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.412147] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1026.412147] env[61911]: value = "task-1251568" [ 1026.412147] env[61911]: _type = "Task" [ 1026.412147] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.422281] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251568, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.497913] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.537609] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba352ad-8292-4632-b0ce-15ed1e27091e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.545532] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac18a3f-4d87-4242-ba32-6de5ddf4a5bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.577450] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395da878-c7e3-4dc7-9836-50ee1860b057 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.584956] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6055c1f-1410-467a-9254-26f0cb9e66c6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.598565] env[61911]: DEBUG nova.compute.provider_tree [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.790377] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.790723] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.790723] env[61911]: DEBUG nova.network.neutron [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.923530] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251568, 'name': Rename_Task, 'duration_secs': 0.205908} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.923836] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.924141] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5a085e8-eccd-40e8-b9e7-82df35bd970e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.930964] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1026.930964] env[61911]: value = "task-1251569" [ 1026.930964] env[61911]: _type = "Task" [ 1026.930964] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.938731] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.103008] env[61911]: DEBUG nova.scheduler.client.report [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1027.336448] env[61911]: DEBUG nova.network.neutron [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.447604] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.547781] env[61911]: DEBUG nova.network.neutron [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19140385-69", "ovs_interfaceid": "19140385-696d-4e4c-a2dd-d53c936d4416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.609810] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.610469] env[61911]: DEBUG nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1027.613566] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.116s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.614565] env[61911]: DEBUG nova.objects.instance [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lazy-loading 'resources' on Instance uuid 31f0a4f7-a69d-4ec4-966d-93f402163b88 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.941658] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.050635] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.050988] env[61911]: DEBUG nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Instance network_info: |[{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19140385-69", "ovs_interfaceid": "19140385-696d-4e4c-a2dd-d53c936d4416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1028.051469] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:3e:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a58387dd-f438-4913-af6a-fafb734cd881', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19140385-696d-4e4c-a2dd-d53c936d4416', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.059256] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating folder: Project (ce8675694c2841a58e87bb7250f3435a). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.059566] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-32c982e0-104c-45e4-bca6-35a39fb97498 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.069705] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created folder: Project (ce8675694c2841a58e87bb7250f3435a) in parent group-v269521. [ 1028.069924] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating folder: Instances. Parent ref: group-v269666. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.070244] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3853aaf-069b-4254-b23d-db5e8cd3a885 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.078541] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created folder: Instances in parent group-v269666. [ 1028.078802] env[61911]: DEBUG oslo.service.loopingcall [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.079009] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.079266] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3b71c5d-bcc0-44c1-9935-fd6deed1d161 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.097747] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.097747] env[61911]: value = "task-1251572" [ 1028.097747] env[61911]: _type = "Task" [ 1028.097747] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.106027] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251572, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.116858] env[61911]: DEBUG nova.compute.utils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1028.121681] env[61911]: DEBUG nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1028.121842] env[61911]: DEBUG nova.network.neutron [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1028.135837] env[61911]: DEBUG nova.compute.manager [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1028.136330] env[61911]: DEBUG nova.compute.manager [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing instance network info cache due to event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1028.136814] env[61911]: DEBUG oslo_concurrency.lockutils [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] Acquiring lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.136904] env[61911]: DEBUG oslo_concurrency.lockutils [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] Acquired lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.137146] env[61911]: DEBUG nova.network.neutron [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.170577] env[61911]: DEBUG nova.policy [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62586445653844078feeeb731bd16f62', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e6ca33655884f16b4c0a6fcdb31ec12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1028.228095] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebca3ecf-7a68-4a99-b475-fed77a78ed74 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.237916] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a08757-60bf-4ba4-9246-95806a2bfc82 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.272990] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc4de89-c380-46d5-8dde-9edfbee01bde {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.285078] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95114f61-587d-4092-b264-ce79e0ec99dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.300448] env[61911]: DEBUG nova.compute.provider_tree [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.444319] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251569, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.577037] env[61911]: DEBUG nova.network.neutron [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Successfully created port: ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.612124] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251572, 'name': CreateVM_Task, 'duration_secs': 0.458958} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.612366] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.613353] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.613609] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.614184] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1028.614588] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52fe49bc-7f8f-4bcc-8b8f-b99ea41b3351 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.621008] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1028.621008] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b64f7-1346-5459-7fe9-09ca5ca7ab02" [ 1028.621008] env[61911]: _type = "Task" [ 1028.621008] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.621876] env[61911]: DEBUG nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1028.637564] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b64f7-1346-5459-7fe9-09ca5ca7ab02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.806876] env[61911]: DEBUG nova.scheduler.client.report [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1028.883099] env[61911]: DEBUG nova.network.neutron [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updated VIF entry in instance network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1028.883644] env[61911]: DEBUG nova.network.neutron [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19140385-69", "ovs_interfaceid": "19140385-696d-4e4c-a2dd-d53c936d4416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.941834] env[61911]: DEBUG oslo_vmware.api [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251569, 'name': PowerOnVM_Task, 'duration_secs': 1.637866} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.942336] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.942397] env[61911]: INFO nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Took 8.80 seconds to spawn the instance on the hypervisor. [ 1028.942574] env[61911]: DEBUG nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1028.943413] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532d8b4a-075f-41d2-8698-7f902a635e4e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.134746] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]522b64f7-1346-5459-7fe9-09ca5ca7ab02, 'name': SearchDatastore_Task, 'duration_secs': 0.030044} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.135335] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.135624] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.135877] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.136049] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.136242] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.136509] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-846c282a-b6ab-4137-a2eb-823d5f92e490 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.144932] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.145152] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.145910] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4196f87-6a62-4b4d-888e-a6faeb71aaf9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.151238] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1029.151238] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52965721-e7d5-8f34-d4ca-0c374a97258f" [ 1029.151238] env[61911]: _type = "Task" [ 1029.151238] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.159578] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52965721-e7d5-8f34-d4ca-0c374a97258f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.313561] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.337995] env[61911]: INFO nova.scheduler.client.report [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Deleted allocations for instance 31f0a4f7-a69d-4ec4-966d-93f402163b88 [ 1029.386717] env[61911]: DEBUG oslo_concurrency.lockutils [req-70e46f29-8cd7-4bef-8b93-2644ff3c8630 req-a7f5aa51-01e2-4981-88ba-398c381fa89b service nova] Releasing lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.459647] env[61911]: INFO nova.compute.manager [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Took 15.25 seconds to build instance. [ 1029.637038] env[61911]: DEBUG nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1029.663660] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52965721-e7d5-8f34-d4ca-0c374a97258f, 'name': SearchDatastore_Task, 'duration_secs': 0.009153} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.664672] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-feda009d-d343-4b16-9d96-37d4720df26c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.670044] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1029.670044] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bb1ee1-54e9-04ff-3a49-96ab7fce7581" [ 1029.670044] env[61911]: _type = "Task" [ 1029.670044] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.678040] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bb1ee1-54e9-04ff-3a49-96ab7fce7581, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.845414] env[61911]: DEBUG oslo_concurrency.lockutils [None req-52a3d950-1773-4e03-bcda-3c6cd3b1f0b5 tempest-ServersTestJSON-168971709 tempest-ServersTestJSON-168971709-project-member] Lock "31f0a4f7-a69d-4ec4-966d-93f402163b88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.779s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.960316] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1029.960803] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.960803] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1029.961021] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.961134] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1029.961885] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1029.961885] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1029.961885] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1029.962091] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1029.962273] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1029.962455] env[61911]: DEBUG nova.virt.hardware [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1029.962926] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4af41c6c-d094-434f-ad2f-99a7bf21f56d tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.761s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.963742] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8811ceb-2729-47ed-b84c-09826cdfa1c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.980823] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3660a263-68a2-4b71-bf29-40cbee661355 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.076782] env[61911]: DEBUG nova.network.neutron [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Successfully updated port: ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1030.110680] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5260464f-9253-b891-7ad5-561e15be94d3/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1030.111868] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37ae6ec-2436-4338-88ee-ccf8dfff00ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.117668] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5260464f-9253-b891-7ad5-561e15be94d3/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1030.117838] env[61911]: ERROR oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5260464f-9253-b891-7ad5-561e15be94d3/disk-0.vmdk due to incomplete transfer. [ 1030.118060] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1914fa1f-04f2-4985-a23b-2db70ceb8f1c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.124197] env[61911]: DEBUG oslo_vmware.rw_handles [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5260464f-9253-b891-7ad5-561e15be94d3/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1030.124393] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Uploaded image dabc0dd7-8ea1-48f4-8766-664d76c32e4a to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1030.126613] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1030.126837] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-802cce6b-1b90-45e2-ac55-c183a6218e26 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.132262] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1030.132262] env[61911]: value = "task-1251573" [ 1030.132262] env[61911]: _type = "Task" [ 1030.132262] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.142811] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251573, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.163460] env[61911]: DEBUG nova.compute.manager [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Received event network-vif-plugged-ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1030.163674] env[61911]: DEBUG oslo_concurrency.lockutils [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] Acquiring lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.163868] env[61911]: DEBUG oslo_concurrency.lockutils [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.164122] env[61911]: DEBUG oslo_concurrency.lockutils [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.164399] env[61911]: DEBUG nova.compute.manager [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] No waiting events found dispatching network-vif-plugged-ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1030.164679] env[61911]: WARNING nova.compute.manager [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Received unexpected event network-vif-plugged-ebc99f07-acb2-44d1-9655-7359b307ecee for instance with vm_state building and task_state spawning. [ 1030.164892] env[61911]: DEBUG nova.compute.manager [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Received event network-changed-ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1030.165106] env[61911]: DEBUG nova.compute.manager [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Refreshing instance network info cache due to event network-changed-ebc99f07-acb2-44d1-9655-7359b307ecee. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1030.165399] env[61911]: DEBUG oslo_concurrency.lockutils [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] Acquiring lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.165613] env[61911]: DEBUG oslo_concurrency.lockutils [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] Acquired lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.165865] env[61911]: DEBUG nova.network.neutron [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Refreshing network info cache for port ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.180153] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52bb1ee1-54e9-04ff-3a49-96ab7fce7581, 'name': SearchDatastore_Task, 'duration_secs': 0.00946} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.180411] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.180660] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.180900] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aff5b607-9559-4d1e-a869-789657c5858c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.186867] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1030.186867] env[61911]: value = "task-1251574" [ 1030.186867] env[61911]: _type = "Task" [ 1030.186867] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.194939] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251574, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.472291] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.475345] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.003s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.475564] env[61911]: INFO nova.compute.manager [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Shelving [ 1030.580493] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.642737] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251573, 'name': Destroy_Task, 'duration_secs': 0.463982} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.643050] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Destroyed the VM [ 1030.643325] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1030.643602] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8441eaee-c818-4c89-89f7-94abcc371c50 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.649420] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1030.649420] env[61911]: value = "task-1251575" [ 1030.649420] env[61911]: _type = "Task" [ 1030.649420] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.657117] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251575, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.695798] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251574, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44686} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.696093] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1030.696323] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1030.696900] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ccbcf85a-33aa-4455-9b21-495228529fe3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.702555] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1030.702555] env[61911]: value = "task-1251576" [ 1030.702555] env[61911]: _type = "Task" [ 1030.702555] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.709676] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251576, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.712524] env[61911]: DEBUG nova.network.neutron [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1030.835552] env[61911]: DEBUG nova.network.neutron [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.160389] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251575, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.214909] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075778} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.215231] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.216069] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8921bc-9048-493b-9980-8eabf5229009 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.240768] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.241137] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d198f42-f93d-47c0-982f-f872d18169dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.262174] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1031.262174] env[61911]: value = "task-1251577" [ 1031.262174] env[61911]: _type = "Task" [ 1031.262174] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.270437] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251577, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.338361] env[61911]: DEBUG oslo_concurrency.lockutils [req-62d2c096-c3ce-4e86-9cc3-ec0b2fa737eb req-ce0a51d9-a005-4afc-95cb-e184c7152e40 service nova] Releasing lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.338761] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.338924] env[61911]: DEBUG nova.network.neutron [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.486864] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.487139] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e5d8f51-7616-4062-b523-e13ccc8337a8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.494271] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1031.494271] env[61911]: value = "task-1251578" [ 1031.494271] env[61911]: _type = "Task" [ 1031.494271] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.502614] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.660892] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251575, 'name': RemoveSnapshot_Task} progress is 43%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.777566] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251577, 'name': ReconfigVM_Task, 'duration_secs': 0.394916} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.777566] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.777566] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67eb4d6f-02c1-44e9-b115-33b66d0416a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.783266] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1031.783266] env[61911]: value = "task-1251579" [ 1031.783266] env[61911]: _type = "Task" [ 1031.783266] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.793651] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251579, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.872702] env[61911]: DEBUG nova.network.neutron [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.006976] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251578, 'name': PowerOffVM_Task, 'duration_secs': 0.195337} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.007408] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.008318] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6eb5ce-01a2-4905-8b03-3ec53e97f481 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.042125] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9630485b-92a7-4341-9443-53b86d51aa5f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.061799] env[61911]: DEBUG nova.network.neutron [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Updating instance_info_cache with network_info: [{"id": "ebc99f07-acb2-44d1-9655-7359b307ecee", "address": "fa:16:3e:1f:16:e2", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc99f07-ac", "ovs_interfaceid": "ebc99f07-acb2-44d1-9655-7359b307ecee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.165217] env[61911]: DEBUG oslo_vmware.api [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251575, 'name': RemoveSnapshot_Task, 'duration_secs': 1.159671} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.165593] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1032.165873] env[61911]: INFO nova.compute.manager [None req-316fd917-50ac-43dd-a20a-6ea9a0907757 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Took 15.39 seconds to snapshot the instance on the hypervisor. [ 1032.292861] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251579, 'name': Rename_Task, 'duration_secs': 0.191794} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.293189] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.293440] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-475c4fd7-6484-4b1c-82e8-9c226c981232 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.304664] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1032.304664] env[61911]: value = "task-1251580" [ 1032.304664] env[61911]: _type = "Task" [ 1032.304664] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.312385] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251580, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.556666] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1032.557032] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3982840a-2bef-4d1f-9646-4efca945362d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.564254] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.564552] env[61911]: DEBUG nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Instance network_info: |[{"id": "ebc99f07-acb2-44d1-9655-7359b307ecee", "address": "fa:16:3e:1f:16:e2", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc99f07-ac", "ovs_interfaceid": "ebc99f07-acb2-44d1-9655-7359b307ecee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1032.564931] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1032.564931] env[61911]: value = "task-1251581" [ 1032.564931] env[61911]: _type = "Task" [ 1032.564931] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.565291] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:16:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ebc99f07-acb2-44d1-9655-7359b307ecee', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1032.572840] env[61911]: DEBUG oslo.service.loopingcall [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.573122] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1032.576574] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-edc48b4e-ac3f-4157-a95a-1b477452f738 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.594815] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251581, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.595942] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1032.595942] env[61911]: value = "task-1251582" [ 1032.595942] env[61911]: _type = "Task" [ 1032.595942] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.603007] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251582, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.815034] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251580, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.081953] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251581, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.105777] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251582, 'name': CreateVM_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.316498] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251580, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.583959] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251581, 'name': CreateSnapshot_Task, 'duration_secs': 0.897622} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.584285] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1033.585053] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a34695e-99e6-4473-a39b-3165d3bd49d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.605019] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251582, 'name': CreateVM_Task, 'duration_secs': 0.619413} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.605192] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1033.605838] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.606029] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.606379] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.606616] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e985e1a-a957-4322-ab39-724162b54833 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.610395] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1033.610395] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529d2b9a-8548-89cc-a510-c689ceb06ff4" [ 1033.610395] env[61911]: _type = "Task" [ 1033.610395] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.617567] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529d2b9a-8548-89cc-a510-c689ceb06ff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.793455] env[61911]: DEBUG nova.compute.manager [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1033.794453] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1f9137-ada2-41e2-afb1-2e6ba978fc5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.815283] env[61911]: DEBUG oslo_vmware.api [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251580, 'name': PowerOnVM_Task, 'duration_secs': 1.160229} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.815532] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1033.815735] env[61911]: INFO nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Took 8.06 seconds to spawn the instance on the hypervisor. [ 1033.815926] env[61911]: DEBUG nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1033.816651] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54762e5-ae79-4da1-9cbf-ce6719dffff7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.102040] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1034.102323] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-22e535f6-c34f-415a-bbd7-f215b0814936 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.111042] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1034.111042] env[61911]: value = "task-1251583" [ 1034.111042] env[61911]: _type = "Task" [ 1034.111042] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.122929] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]529d2b9a-8548-89cc-a510-c689ceb06ff4, 'name': SearchDatastore_Task, 'duration_secs': 0.009554} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.125964] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.126248] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.126491] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.126647] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.126834] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.127093] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251583, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.127293] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef067512-af06-4090-aa81-99b754d0acb5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.134478] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.134672] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.135347] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdd5a6b5-011f-4a31-bdef-44be35eba3be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.139583] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1034.139583] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5219f22a-7efb-79ed-12fe-3b50c807e38b" [ 1034.139583] env[61911]: _type = "Task" [ 1034.139583] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.146580] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5219f22a-7efb-79ed-12fe-3b50c807e38b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.306054] env[61911]: INFO nova.compute.manager [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] instance snapshotting [ 1034.309572] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3550d18-995d-4ad9-8a7e-347cd78f1107 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.343030] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e278c9-bc31-45a2-bf71-5ea047cd90cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.349307] env[61911]: INFO nova.compute.manager [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Took 16.11 seconds to build instance. [ 1034.624124] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251583, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.650177] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5219f22a-7efb-79ed-12fe-3b50c807e38b, 'name': SearchDatastore_Task, 'duration_secs': 0.007795} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.651021] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d98a8dcc-1c69-4ea6-8873-3999d8f0c455 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.656197] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1034.656197] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524b95ec-b03b-a2bd-132f-d73270ab99e4" [ 1034.656197] env[61911]: _type = "Task" [ 1034.656197] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.663631] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524b95ec-b03b-a2bd-132f-d73270ab99e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.857210] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9ce94dfa-603a-4945-865f-f1e2f1a90d5f tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.628s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.858262] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1034.858684] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3c9d6a56-e58e-421d-abd4-86b2d1d4aadb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.866304] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1034.866304] env[61911]: value = "task-1251584" [ 1034.866304] env[61911]: _type = "Task" [ 1034.866304] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.874511] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251584, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.050289] env[61911]: DEBUG nova.compute.manager [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1035.050496] env[61911]: DEBUG nova.compute.manager [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing instance network info cache due to event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1035.050726] env[61911]: DEBUG oslo_concurrency.lockutils [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] Acquiring lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.050878] env[61911]: DEBUG oslo_concurrency.lockutils [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] Acquired lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.051214] env[61911]: DEBUG nova.network.neutron [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1035.126588] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251583, 'name': CloneVM_Task} progress is 95%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.170109] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524b95ec-b03b-a2bd-132f-d73270ab99e4, 'name': SearchDatastore_Task, 'duration_secs': 0.008896} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.170445] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.170716] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0/8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1035.170992] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14cf407c-2a88-423d-bc3a-5c0bc543be02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.177354] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1035.177354] env[61911]: value = "task-1251585" [ 1035.177354] env[61911]: _type = "Task" [ 1035.177354] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.187262] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251585, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.377424] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251584, 'name': CreateSnapshot_Task, 'duration_secs': 0.420201} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.378029] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1035.378479] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c86160-6430-464d-918d-d460e4fb29ee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.625111] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251583, 'name': CloneVM_Task, 'duration_secs': 1.362533} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.625411] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Created linked-clone VM from snapshot [ 1035.626196] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15eb2b6-cdf6-4370-8d3f-a744a610c3a5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.635677] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Uploading image 7360855b-31b5-45dc-bc37-f8a6c23c77a0 {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1035.659405] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1035.659405] env[61911]: value = "vm-269671" [ 1035.659405] env[61911]: _type = "VirtualMachine" [ 1035.659405] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1035.659706] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d6a24889-06e4-47c9-8ca7-1b895a17beeb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.669354] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lease: (returnval){ [ 1035.669354] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52877a9d-abfa-de04-1b16-9bc13be4855b" [ 1035.669354] env[61911]: _type = "HttpNfcLease" [ 1035.669354] env[61911]: } obtained for exporting VM: (result){ [ 1035.669354] env[61911]: value = "vm-269671" [ 1035.669354] env[61911]: _type = "VirtualMachine" [ 1035.669354] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1035.669615] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the lease: (returnval){ [ 1035.669615] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52877a9d-abfa-de04-1b16-9bc13be4855b" [ 1035.669615] env[61911]: _type = "HttpNfcLease" [ 1035.669615] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1035.675980] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1035.675980] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52877a9d-abfa-de04-1b16-9bc13be4855b" [ 1035.675980] env[61911]: _type = "HttpNfcLease" [ 1035.675980] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1035.688098] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251585, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42941} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.688358] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0/8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.688575] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.688825] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b5f9df8-4cbc-4460-a701-a636fe016e1a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.694023] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1035.694023] env[61911]: value = "task-1251587" [ 1035.694023] env[61911]: _type = "Task" [ 1035.694023] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.701682] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.783715] env[61911]: DEBUG nova.network.neutron [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updated VIF entry in instance network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1035.784181] env[61911]: DEBUG nova.network.neutron [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19140385-69", "ovs_interfaceid": "19140385-696d-4e4c-a2dd-d53c936d4416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.897981] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1035.898301] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-62769869-5128-4541-9789-c5ec3e8d57ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.907385] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1035.907385] env[61911]: value = "task-1251588" [ 1035.907385] env[61911]: _type = "Task" [ 1035.907385] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.915608] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251588, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.178586] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1036.178586] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52877a9d-abfa-de04-1b16-9bc13be4855b" [ 1036.178586] env[61911]: _type = "HttpNfcLease" [ 1036.178586] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1036.178917] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1036.178917] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52877a9d-abfa-de04-1b16-9bc13be4855b" [ 1036.178917] env[61911]: _type = "HttpNfcLease" [ 1036.178917] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1036.179668] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d686e688-87b6-4da6-8569-4cd27689ac9b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.186865] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52aaeef7-201c-e11f-0863-32bd5f18bcda/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1036.187066] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52aaeef7-201c-e11f-0863-32bd5f18bcda/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1036.253099] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062301} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.253358] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.254295] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cde0705-a30f-4fdf-8f15-e952e81a8381 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.277250] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0/8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.277445] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e36d5d2d-3a7d-40f3-ae70-88ef69b7acfd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.291681] env[61911]: DEBUG oslo_concurrency.lockutils [req-dfbb07aa-5326-45b8-8db4-13468ba8a547 req-cc758453-4c6f-4fe6-abd5-f1e7a9541a11 service nova] Releasing lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.298490] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1036.298490] env[61911]: value = "task-1251589" [ 1036.298490] env[61911]: _type = "Task" [ 1036.298490] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.308185] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.415032] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-571e392f-9f90-48a8-99d0-267cd2562efe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.419929] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251588, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.809083] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251589, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.918575] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251588, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.309467] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251589, 'name': ReconfigVM_Task, 'duration_secs': 0.685257} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.309855] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0/8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.310561] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04f7d5e8-30d7-4b34-bbeb-27b3f3074d98 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.317096] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1037.317096] env[61911]: value = "task-1251590" [ 1037.317096] env[61911]: _type = "Task" [ 1037.317096] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.327083] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251590, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.418793] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251588, 'name': CloneVM_Task, 'duration_secs': 1.424787} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.419288] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Created linked-clone VM from snapshot [ 1037.420102] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3987cd9-dfd1-4fd0-89a0-314c5c1eb6d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.428142] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Uploading image e4c53c7d-1fa3-4ba3-8852-0e6606e8d910 {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1037.450385] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1037.450385] env[61911]: value = "vm-269673" [ 1037.450385] env[61911]: _type = "VirtualMachine" [ 1037.450385] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1037.450833] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bf700258-fcc2-48bb-be8b-6367db841342 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.456821] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lease: (returnval){ [ 1037.456821] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280651c-d047-43f8-7aad-5146084a03c8" [ 1037.456821] env[61911]: _type = "HttpNfcLease" [ 1037.456821] env[61911]: } obtained for exporting VM: (result){ [ 1037.456821] env[61911]: value = "vm-269673" [ 1037.456821] env[61911]: _type = "VirtualMachine" [ 1037.456821] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1037.457082] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the lease: (returnval){ [ 1037.457082] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280651c-d047-43f8-7aad-5146084a03c8" [ 1037.457082] env[61911]: _type = "HttpNfcLease" [ 1037.457082] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1037.463207] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1037.463207] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280651c-d047-43f8-7aad-5146084a03c8" [ 1037.463207] env[61911]: _type = "HttpNfcLease" [ 1037.463207] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1037.827356] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251590, 'name': Rename_Task, 'duration_secs': 0.230872} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.827647] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.828156] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edc10db8-9338-4bfa-ae7d-75cbd993f3a5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.833946] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1037.833946] env[61911]: value = "task-1251592" [ 1037.833946] env[61911]: _type = "Task" [ 1037.833946] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.841303] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251592, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.965936] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1037.965936] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280651c-d047-43f8-7aad-5146084a03c8" [ 1037.965936] env[61911]: _type = "HttpNfcLease" [ 1037.965936] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1037.966264] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1037.966264] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5280651c-d047-43f8-7aad-5146084a03c8" [ 1037.966264] env[61911]: _type = "HttpNfcLease" [ 1037.966264] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1037.967037] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbdd284-92f1-4ee2-a240-8a4594f17ea5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.974176] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523df81d-3c6d-6287-609c-88c418baab14/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1037.974429] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523df81d-3c6d-6287-609c-88c418baab14/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1038.069438] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2ba825cd-6b35-41b0-9dea-70d284e5f826 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.345251] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251592, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.844690] env[61911]: DEBUG oslo_vmware.api [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251592, 'name': PowerOnVM_Task, 'duration_secs': 0.688144} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.845213] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.845561] env[61911]: INFO nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Took 9.21 seconds to spawn the instance on the hypervisor. [ 1038.845907] env[61911]: DEBUG nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1038.846758] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e68bb59-4dcf-4e2f-a3cc-4f02737a8c79 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.365629] env[61911]: INFO nova.compute.manager [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Took 15.79 seconds to build instance. [ 1039.867734] env[61911]: DEBUG oslo_concurrency.lockutils [None req-af2537a3-6d91-42b8-be85-c15424484b9a tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.303s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.144521] env[61911]: DEBUG nova.compute.manager [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Received event network-changed-ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1040.145388] env[61911]: DEBUG nova.compute.manager [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Refreshing instance network info cache due to event network-changed-ebc99f07-acb2-44d1-9655-7359b307ecee. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1040.145675] env[61911]: DEBUG oslo_concurrency.lockutils [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] Acquiring lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.145799] env[61911]: DEBUG oslo_concurrency.lockutils [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] Acquired lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.145974] env[61911]: DEBUG nova.network.neutron [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Refreshing network info cache for port ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.883388] env[61911]: DEBUG nova.network.neutron [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Updated VIF entry in instance network info cache for port ebc99f07-acb2-44d1-9655-7359b307ecee. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1040.883865] env[61911]: DEBUG nova.network.neutron [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Updating instance_info_cache with network_info: [{"id": "ebc99f07-acb2-44d1-9655-7359b307ecee", "address": "fa:16:3e:1f:16:e2", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebc99f07-ac", "ovs_interfaceid": "ebc99f07-acb2-44d1-9655-7359b307ecee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.387029] env[61911]: DEBUG oslo_concurrency.lockutils [req-fac925fc-b505-4d47-9767-c4008caad840 req-0d1a0b92-8b86-439c-a725-141737a0dbb9 service nova] Releasing lock "refresh_cache-8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.775125] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52aaeef7-201c-e11f-0863-32bd5f18bcda/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1044.775745] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae8f19c-2256-4abf-b0ef-647d71e6f8a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.783508] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52aaeef7-201c-e11f-0863-32bd5f18bcda/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1044.783721] env[61911]: ERROR oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52aaeef7-201c-e11f-0863-32bd5f18bcda/disk-0.vmdk due to incomplete transfer. [ 1044.783969] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-afd0cfbd-efab-4676-b25e-8f20d44c4da4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.791581] env[61911]: DEBUG oslo_vmware.rw_handles [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52aaeef7-201c-e11f-0863-32bd5f18bcda/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1044.791795] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Uploaded image 7360855b-31b5-45dc-bc37-f8a6c23c77a0 to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1044.794268] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1044.794549] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8659cba3-7a40-4ce4-8a21-81aa79ca0c41 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.801195] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1044.801195] env[61911]: value = "task-1251593" [ 1044.801195] env[61911]: _type = "Task" [ 1044.801195] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.810020] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251593, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.310980] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251593, 'name': Destroy_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.362816] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523df81d-3c6d-6287-609c-88c418baab14/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1045.363769] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528b9faa-b34a-4ee7-8f68-bafbcf75e79e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.370391] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523df81d-3c6d-6287-609c-88c418baab14/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1045.370565] env[61911]: ERROR oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523df81d-3c6d-6287-609c-88c418baab14/disk-0.vmdk due to incomplete transfer. [ 1045.370779] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d57ee8e8-af58-4bc0-87f4-c843edcddaaf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.378306] env[61911]: DEBUG oslo_vmware.rw_handles [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523df81d-3c6d-6287-609c-88c418baab14/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1045.378498] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Uploaded image e4c53c7d-1fa3-4ba3-8852-0e6606e8d910 to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1045.380279] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1045.380511] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-61d939ea-f1b8-4be9-b464-63a5a0c385cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.385754] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1045.385754] env[61911]: value = "task-1251594" [ 1045.385754] env[61911]: _type = "Task" [ 1045.385754] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.393257] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251594, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.812321] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251593, 'name': Destroy_Task} progress is 33%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.895014] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251594, 'name': Destroy_Task, 'duration_secs': 0.385603} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.895384] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Destroyed the VM [ 1045.895630] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1045.895871] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bff47810-e294-42d1-ba50-1de09b0808d4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.901342] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1045.901342] env[61911]: value = "task-1251595" [ 1045.901342] env[61911]: _type = "Task" [ 1045.901342] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.908693] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251595, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.312997] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251593, 'name': Destroy_Task, 'duration_secs': 1.081283} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.313292] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Destroyed the VM [ 1046.313560] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1046.313825] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b6a79e64-55f2-43b3-a798-bc0539dd8135 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.319696] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1046.319696] env[61911]: value = "task-1251596" [ 1046.319696] env[61911]: _type = "Task" [ 1046.319696] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.327213] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251596, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.412432] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251595, 'name': RemoveSnapshot_Task} progress is 64%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.829844] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251596, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.913164] env[61911]: DEBUG oslo_vmware.api [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251595, 'name': RemoveSnapshot_Task, 'duration_secs': 0.550319} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.913480] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1046.913725] env[61911]: INFO nova.compute.manager [None req-bf5a1e86-fdd3-443d-a9d7-9df40781a43b tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Took 12.61 seconds to snapshot the instance on the hypervisor. [ 1047.331454] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251596, 'name': RemoveSnapshot_Task, 'duration_secs': 0.543056} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.331744] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1047.332278] env[61911]: DEBUG nova.compute.manager [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1047.332894] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9c2799-020e-4610-8307-801e0a8bb37f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.845213] env[61911]: INFO nova.compute.manager [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Shelve offloading [ 1048.348711] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.349079] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-097667ef-db1e-4e5a-95c2-b4bf13ac15b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.356258] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1048.356258] env[61911]: value = "task-1251597" [ 1048.356258] env[61911]: _type = "Task" [ 1048.356258] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.363976] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251597, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.666964] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "632e811e-b253-47ed-9d1e-6b2538ff804b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.666964] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.667171] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "632e811e-b253-47ed-9d1e-6b2538ff804b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.667365] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.667543] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.669819] env[61911]: INFO nova.compute.manager [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Terminating instance [ 1048.866173] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1048.866470] env[61911]: DEBUG nova.compute.manager [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1048.867128] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e00a0a-58f8-419d-9e5a-798ba477d75b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.872421] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.872587] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.872758] env[61911]: DEBUG nova.network.neutron [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.174124] env[61911]: DEBUG nova.compute.manager [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1049.174500] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.175904] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dd701a-9a99-4609-ba5f-805afa7547d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.186802] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.187156] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3275f4d7-d0bf-432f-b676-6b8f9f191496 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.194519] env[61911]: DEBUG oslo_vmware.api [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1049.194519] env[61911]: value = "task-1251598" [ 1049.194519] env[61911]: _type = "Task" [ 1049.194519] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.206242] env[61911]: DEBUG oslo_vmware.api [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.704519] env[61911]: DEBUG oslo_vmware.api [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251598, 'name': PowerOffVM_Task, 'duration_secs': 0.217019} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.704808] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.704984] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.705242] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ada821e9-c26c-4250-af3f-663f5fd33061 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.721215] env[61911]: DEBUG nova.network.neutron [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Updating instance_info_cache with network_info: [{"id": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "address": "fa:16:3e:58:3a:db", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b158ae-a6", "ovs_interfaceid": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.762679] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.762895] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.763113] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Deleting the datastore file [datastore2] 632e811e-b253-47ed-9d1e-6b2538ff804b {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.763729] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d835c88-8db1-4dc3-a0bf-ed6e76b553a0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.769897] env[61911]: DEBUG oslo_vmware.api [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1049.769897] env[61911]: value = "task-1251600" [ 1049.769897] env[61911]: _type = "Task" [ 1049.769897] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.777503] env[61911]: DEBUG oslo_vmware.api [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.223982] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.279880] env[61911]: DEBUG oslo_vmware.api [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251600, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14806} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.280173] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.280367] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.280550] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.280732] env[61911]: INFO nova.compute.manager [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1050.280971] env[61911]: DEBUG oslo.service.loopingcall [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.281215] env[61911]: DEBUG nova.compute.manager [-] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1050.281320] env[61911]: DEBUG nova.network.neutron [-] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1050.461854] env[61911]: DEBUG nova.compute.manager [req-907807f4-4613-45ba-ac92-f1cfc3813710 req-74870234-e0e3-4ecd-86b3-2f6b5f6e5f8a service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Received event network-vif-unplugged-92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1050.462110] env[61911]: DEBUG oslo_concurrency.lockutils [req-907807f4-4613-45ba-ac92-f1cfc3813710 req-74870234-e0e3-4ecd-86b3-2f6b5f6e5f8a service nova] Acquiring lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.462299] env[61911]: DEBUG oslo_concurrency.lockutils [req-907807f4-4613-45ba-ac92-f1cfc3813710 req-74870234-e0e3-4ecd-86b3-2f6b5f6e5f8a service nova] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.462495] env[61911]: DEBUG oslo_concurrency.lockutils [req-907807f4-4613-45ba-ac92-f1cfc3813710 req-74870234-e0e3-4ecd-86b3-2f6b5f6e5f8a service nova] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.462637] env[61911]: DEBUG nova.compute.manager [req-907807f4-4613-45ba-ac92-f1cfc3813710 req-74870234-e0e3-4ecd-86b3-2f6b5f6e5f8a service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] No waiting events found dispatching network-vif-unplugged-92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1050.462850] env[61911]: WARNING nova.compute.manager [req-907807f4-4613-45ba-ac92-f1cfc3813710 req-74870234-e0e3-4ecd-86b3-2f6b5f6e5f8a service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Received unexpected event network-vif-unplugged-92b158ae-a674-4ba7-a4a5-c4819bba4492 for instance with vm_state shelved and task_state shelving_offloading. [ 1050.485828] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1050.486831] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8478bb-9416-43d3-8f73-17910b2a40cc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.494782] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.495037] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f04f7ff-4f6e-4194-9159-5c248b49cd69 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.544733] env[61911]: DEBUG nova.compute.manager [req-900cb6c7-835b-4690-9740-139a8d796370 req-34dfb396-6d95-4e7a-b89b-4b03f90655d9 service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Received event network-vif-deleted-07ff77e7-946d-4de9-8e85-badc55c396c1 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1050.544998] env[61911]: INFO nova.compute.manager [req-900cb6c7-835b-4690-9740-139a8d796370 req-34dfb396-6d95-4e7a-b89b-4b03f90655d9 service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Neutron deleted interface 07ff77e7-946d-4de9-8e85-badc55c396c1; detaching it from the instance and deleting it from the info cache [ 1050.545135] env[61911]: DEBUG nova.network.neutron [req-900cb6c7-835b-4690-9740-139a8d796370 req-34dfb396-6d95-4e7a-b89b-4b03f90655d9 service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.555691] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.555911] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.556116] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleting the datastore file [datastore2] d4e33bff-8d26-49de-98c8-d9dcbfc640dd {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.556624] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-437d3116-42a7-4ea1-963c-822c567b188f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.564298] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1050.564298] env[61911]: value = "task-1251602" [ 1050.564298] env[61911]: _type = "Task" [ 1050.564298] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.572267] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.013973] env[61911]: DEBUG nova.network.neutron [-] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.047634] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f81ca14a-43c2-4340-bcf1-357b8925274d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.056796] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57069695-927e-4720-9164-1c5be04de6ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.074385] env[61911]: DEBUG oslo_vmware.api [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134868} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.074633] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.074821] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1051.075007] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1051.082181] env[61911]: DEBUG nova.compute.manager [req-900cb6c7-835b-4690-9740-139a8d796370 req-34dfb396-6d95-4e7a-b89b-4b03f90655d9 service nova] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Detach interface failed, port_id=07ff77e7-946d-4de9-8e85-badc55c396c1, reason: Instance 632e811e-b253-47ed-9d1e-6b2538ff804b could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1051.098318] env[61911]: INFO nova.scheduler.client.report [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted allocations for instance d4e33bff-8d26-49de-98c8-d9dcbfc640dd [ 1051.517246] env[61911]: INFO nova.compute.manager [-] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Took 1.24 seconds to deallocate network for instance. [ 1051.603990] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.603990] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.603990] env[61911]: DEBUG nova.objects.instance [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'resources' on Instance uuid d4e33bff-8d26-49de-98c8-d9dcbfc640dd {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.023456] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.105820] env[61911]: DEBUG nova.objects.instance [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'numa_topology' on Instance uuid d4e33bff-8d26-49de-98c8-d9dcbfc640dd {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.345762] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.487229] env[61911]: DEBUG nova.compute.manager [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Received event network-changed-92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1052.487419] env[61911]: DEBUG nova.compute.manager [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Refreshing instance network info cache due to event network-changed-92b158ae-a674-4ba7-a4a5-c4819bba4492. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1052.487642] env[61911]: DEBUG oslo_concurrency.lockutils [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] Acquiring lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.487791] env[61911]: DEBUG oslo_concurrency.lockutils [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] Acquired lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.487961] env[61911]: DEBUG nova.network.neutron [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Refreshing network info cache for port 92b158ae-a674-4ba7-a4a5-c4819bba4492 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1052.607863] env[61911]: DEBUG nova.objects.base [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1052.668592] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c383f9c-cf26-42ef-984e-be12872b0da4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.676085] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b329ff71-fe53-433e-9a7b-4df4df9e46f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.704707] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d81a8be-f08b-456e-934f-979e52eff578 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.711570] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222215ef-2c04-40db-a22d-cdcc3843e7e9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.725551] env[61911]: DEBUG nova.compute.provider_tree [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.217950] env[61911]: DEBUG nova.network.neutron [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Updated VIF entry in instance network info cache for port 92b158ae-a674-4ba7-a4a5-c4819bba4492. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1053.218356] env[61911]: DEBUG nova.network.neutron [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Updating instance_info_cache with network_info: [{"id": "92b158ae-a674-4ba7-a4a5-c4819bba4492", "address": "fa:16:3e:58:3a:db", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": null, "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap92b158ae-a6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.228649] env[61911]: DEBUG nova.scheduler.client.report [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1053.721409] env[61911]: DEBUG oslo_concurrency.lockutils [req-6a0d9824-a610-4080-a9ed-55e1ddc1f2ea req-ac0f9b89-ef23-430d-b595-e7ca36c15af3 service nova] Releasing lock "refresh_cache-d4e33bff-8d26-49de-98c8-d9dcbfc640dd" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.733489] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.130s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.735877] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.713s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.736118] env[61911]: DEBUG nova.objects.instance [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lazy-loading 'resources' on Instance uuid 632e811e-b253-47ed-9d1e-6b2538ff804b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.246455] env[61911]: DEBUG oslo_concurrency.lockutils [None req-61bda2c7-b2b2-4719-9407-640e71c60e71 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.771s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.247387] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.902s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.247608] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.247818] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.247992] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.249985] env[61911]: INFO nova.compute.manager [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Terminating instance [ 1054.298031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18b0649-4ecc-4af6-b951-e53d6e0a5134 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.304703] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba29db15-e35f-49c5-acc1-43be10ee7c51 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.333557] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eec00f6-5e10-4a16-9f87-553f2c86b478 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.340168] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934ba27e-59d5-49f5-87e7-005a87856d34 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.354081] env[61911]: DEBUG nova.compute.provider_tree [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.594665] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.594925] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.595093] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1054.753563] env[61911]: DEBUG nova.compute.manager [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1054.753818] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.754148] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8db4c9cf-4d70-49a8-a7f7-9caa44d1f498 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.763549] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99e2ca6-5baf-45bb-a0c3-bc21990d04b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.786613] env[61911]: WARNING nova.virt.vmwareapi.vmops [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d4e33bff-8d26-49de-98c8-d9dcbfc640dd could not be found. [ 1054.786818] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1054.787014] env[61911]: INFO nova.compute.manager [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 1054.787269] env[61911]: DEBUG oslo.service.loopingcall [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.787491] env[61911]: DEBUG nova.compute.manager [-] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1054.787586] env[61911]: DEBUG nova.network.neutron [-] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1054.857012] env[61911]: DEBUG nova.scheduler.client.report [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1055.362142] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.382332] env[61911]: INFO nova.scheduler.client.report [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Deleted allocations for instance 632e811e-b253-47ed-9d1e-6b2538ff804b [ 1055.475863] env[61911]: DEBUG nova.network.neutron [-] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.889614] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0af00190-6613-4291-ae82-d96e48404d04 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "632e811e-b253-47ed-9d1e-6b2538ff804b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.222s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.978533] env[61911]: INFO nova.compute.manager [-] [instance: d4e33bff-8d26-49de-98c8-d9dcbfc640dd] Took 1.19 seconds to deallocate network for instance. [ 1056.362024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "b594227d-1647-4e43-8c90-962a9e765388" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.362024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "b594227d-1647-4e43-8c90-962a9e765388" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.362208] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "b594227d-1647-4e43-8c90-962a9e765388-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.362343] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "b594227d-1647-4e43-8c90-962a9e765388-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.362494] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "b594227d-1647-4e43-8c90-962a9e765388-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.364790] env[61911]: INFO nova.compute.manager [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Terminating instance [ 1056.868901] env[61911]: DEBUG nova.compute.manager [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1056.869170] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.870100] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f38efb5-4f5f-4bc5-9a76-ede38ee20f73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.878442] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.878674] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec877965-d624-4a67-b200-8d832a08ea08 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.885275] env[61911]: DEBUG oslo_vmware.api [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1056.885275] env[61911]: value = "task-1251603" [ 1056.885275] env[61911]: _type = "Task" [ 1056.885275] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.892821] env[61911]: DEBUG oslo_vmware.api [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.003612] env[61911]: DEBUG oslo_concurrency.lockutils [None req-99cc07d1-24ce-434b-8021-7825abfc9d40 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "d4e33bff-8d26-49de-98c8-d9dcbfc640dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.756s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.395484] env[61911]: DEBUG oslo_vmware.api [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251603, 'name': PowerOffVM_Task, 'duration_secs': 0.184787} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.395766] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.395958] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.396428] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b74efc3-08f5-492d-9ae6-3aed00c88b4b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.494058] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.494277] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.494481] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Deleting the datastore file [datastore1] b594227d-1647-4e43-8c90-962a9e765388 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.494752] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4be22b74-4755-43a1-be75-ac0ed725bb30 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.500560] env[61911]: DEBUG oslo_vmware.api [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for the task: (returnval){ [ 1057.500560] env[61911]: value = "task-1251605" [ 1057.500560] env[61911]: _type = "Task" [ 1057.500560] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.508255] env[61911]: DEBUG oslo_vmware.api [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.010512] env[61911]: DEBUG oslo_vmware.api [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Task: {'id': task-1251605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140532} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.010917] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.010978] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1058.011153] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1058.011344] env[61911]: INFO nova.compute.manager [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] [instance: b594227d-1647-4e43-8c90-962a9e765388] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1058.011590] env[61911]: DEBUG oslo.service.loopingcall [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.011788] env[61911]: DEBUG nova.compute.manager [-] [instance: b594227d-1647-4e43-8c90-962a9e765388] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1058.011887] env[61911]: DEBUG nova.network.neutron [-] [instance: b594227d-1647-4e43-8c90-962a9e765388] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1058.277751] env[61911]: DEBUG nova.compute.manager [req-02d17c86-d383-4311-b29c-3e9a5039ad03 req-9cfc89da-66de-41d5-9bc2-7053514ef302 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Received event network-vif-deleted-681e329b-e012-42f2-bb1c-d1c63369e251 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1058.278017] env[61911]: INFO nova.compute.manager [req-02d17c86-d383-4311-b29c-3e9a5039ad03 req-9cfc89da-66de-41d5-9bc2-7053514ef302 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Neutron deleted interface 681e329b-e012-42f2-bb1c-d1c63369e251; detaching it from the instance and deleting it from the info cache [ 1058.278355] env[61911]: DEBUG nova.network.neutron [req-02d17c86-d383-4311-b29c-3e9a5039ad03 req-9cfc89da-66de-41d5-9bc2-7053514ef302 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.316328] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "14986d4b-77af-461e-949e-fbc15f4ea6da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.316632] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.757342] env[61911]: DEBUG nova.network.neutron [-] [instance: b594227d-1647-4e43-8c90-962a9e765388] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.781421] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44f0d626-a849-4577-84ae-e5b576d87263 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.791520] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dbb696-347f-4556-9f36-1ee185c37c27 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.814676] env[61911]: DEBUG nova.compute.manager [req-02d17c86-d383-4311-b29c-3e9a5039ad03 req-9cfc89da-66de-41d5-9bc2-7053514ef302 service nova] [instance: b594227d-1647-4e43-8c90-962a9e765388] Detach interface failed, port_id=681e329b-e012-42f2-bb1c-d1c63369e251, reason: Instance b594227d-1647-4e43-8c90-962a9e765388 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1058.818279] env[61911]: DEBUG nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1059.259823] env[61911]: INFO nova.compute.manager [-] [instance: b594227d-1647-4e43-8c90-962a9e765388] Took 1.25 seconds to deallocate network for instance. [ 1059.338085] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.338365] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.339842] env[61911]: INFO nova.compute.claims [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.766954] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.125858] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Didn't find any instances for network info cache update. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 1060.126185] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.126451] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.126671] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.126893] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.127133] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.127357] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.127547] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1060.127758] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.403785] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f32f50-4481-4c30-a1c5-bf1ca1871df5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.411512] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7cee777-b995-47af-8fa6-bceaf8830c8c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.440298] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700a0cef-0cd5-43a7-94d1-746bf9bbe29f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.447201] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1846e0d3-7ff3-4555-9c99-839f244427d3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.459653] env[61911]: DEBUG nova.compute.provider_tree [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.631099] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.962606] env[61911]: DEBUG nova.scheduler.client.report [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1061.468597] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.130s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.469197] env[61911]: DEBUG nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1061.471728] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.705s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.472021] env[61911]: DEBUG nova.objects.instance [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lazy-loading 'resources' on Instance uuid b594227d-1647-4e43-8c90-962a9e765388 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.974797] env[61911]: DEBUG nova.compute.utils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1061.978511] env[61911]: DEBUG nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1061.978685] env[61911]: DEBUG nova.network.neutron [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1062.026147] env[61911]: DEBUG nova.policy [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a5252f9309f4204a3a3fd7d865db6e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b285793da304b5e8f9736098ef163c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1062.039039] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092e1600-fe46-4204-b4f7-6c0ddc8ba53c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.046303] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523d5ca3-c75f-4c15-8e7e-35b74ad07ac7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.076656] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f2c1cd-aa63-4d85-9dfc-a33999c9e54a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.083739] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c427b66b-bfb6-4ed1-b960-b326372fe0ec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.096520] env[61911]: DEBUG nova.compute.provider_tree [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.265277] env[61911]: DEBUG nova.network.neutron [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Successfully created port: 2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.482007] env[61911]: DEBUG nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1062.600059] env[61911]: DEBUG nova.scheduler.client.report [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1063.106045] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.107512] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.477s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.107707] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.108086] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1063.108749] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7d2aff-1a71-40d7-9401-d4a26a26c847 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.116919] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1623a119-cc86-4768-ad6d-a1080c8d3fb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.132314] env[61911]: INFO nova.scheduler.client.report [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Deleted allocations for instance b594227d-1647-4e43-8c90-962a9e765388 [ 1063.134952] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c4337d-eb7b-4b62-ada9-446130fbeeb6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.142520] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be97cf3-ddb3-4af5-bfcd-639b6cf7f0d8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.171149] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180847MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1063.171312] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.171501] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.490767] env[61911]: DEBUG nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1063.515715] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1063.515933] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.516116] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1063.516407] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.516465] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1063.516608] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1063.516821] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1063.516989] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1063.517178] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1063.517372] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1063.517582] env[61911]: DEBUG nova.virt.hardware [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1063.518479] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7642ad1a-0414-46f3-bef7-5e178c2f88bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.526776] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22b8188-7be0-4da4-bd25-788f7539a35a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.644312] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f4a30e4e-902c-4cc6-8dac-688cb3deae83 tempest-ListImageFiltersTestJSON-1925238576 tempest-ListImageFiltersTestJSON-1925238576-project-member] Lock "b594227d-1647-4e43-8c90-962a9e765388" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.282s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.655419] env[61911]: DEBUG nova.compute.manager [req-32350bb7-d08f-411c-b1c0-86bee0c49108 req-f8a117fc-0eb9-4d6f-a791-c1f3839981f2 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Received event network-vif-plugged-2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1063.655419] env[61911]: DEBUG oslo_concurrency.lockutils [req-32350bb7-d08f-411c-b1c0-86bee0c49108 req-f8a117fc-0eb9-4d6f-a791-c1f3839981f2 service nova] Acquiring lock "14986d4b-77af-461e-949e-fbc15f4ea6da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.655533] env[61911]: DEBUG oslo_concurrency.lockutils [req-32350bb7-d08f-411c-b1c0-86bee0c49108 req-f8a117fc-0eb9-4d6f-a791-c1f3839981f2 service nova] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.655651] env[61911]: DEBUG oslo_concurrency.lockutils [req-32350bb7-d08f-411c-b1c0-86bee0c49108 req-f8a117fc-0eb9-4d6f-a791-c1f3839981f2 service nova] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.655844] env[61911]: DEBUG nova.compute.manager [req-32350bb7-d08f-411c-b1c0-86bee0c49108 req-f8a117fc-0eb9-4d6f-a791-c1f3839981f2 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] No waiting events found dispatching network-vif-plugged-2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1063.656038] env[61911]: WARNING nova.compute.manager [req-32350bb7-d08f-411c-b1c0-86bee0c49108 req-f8a117fc-0eb9-4d6f-a791-c1f3839981f2 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Received unexpected event network-vif-plugged-2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f for instance with vm_state building and task_state spawning. [ 1063.746737] env[61911]: DEBUG nova.network.neutron [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Successfully updated port: 2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.195987] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 95853730-1908-42f4-b30f-a4f276d2da8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.196076] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.196153] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 14986d4b-77af-461e-949e-fbc15f4ea6da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.196427] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1064.196468] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1064.242952] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a629a12-12d5-4f39-8056-d1e3fdc2c7b9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.250729] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169cc8e3-850b-43d8-87a9-9fffd8d0345c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.253986] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-14986d4b-77af-461e-949e-fbc15f4ea6da" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.254108] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-14986d4b-77af-461e-949e-fbc15f4ea6da" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.254265] env[61911]: DEBUG nova.network.neutron [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1064.283061] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da2892b-6289-41ad-863a-f267f0962cb7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.290713] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10edbfc2-a296-4d19-9d20-a64d915af630 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.304098] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.789138] env[61911]: DEBUG nova.network.neutron [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1064.807582] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1064.948363] env[61911]: DEBUG nova.network.neutron [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Updating instance_info_cache with network_info: [{"id": "2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f", "address": "fa:16:3e:54:ee:05", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bbe32e9-4c", "ovs_interfaceid": "2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.271594] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "1198724e-2802-47a0-a7e8-3941e3ffc175" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.271829] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "1198724e-2802-47a0-a7e8-3941e3ffc175" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.313329] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1065.313518] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.142s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.450451] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.450719] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.452384] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-14986d4b-77af-461e-949e-fbc15f4ea6da" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.452677] env[61911]: DEBUG nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Instance network_info: |[{"id": "2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f", "address": "fa:16:3e:54:ee:05", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bbe32e9-4c", "ovs_interfaceid": "2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1065.453305] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:ee:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.461040] env[61911]: DEBUG oslo.service.loopingcall [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.462062] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1065.462330] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c9270e8-f6c1-4c99-9784-684312472ed2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.483024] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.483024] env[61911]: value = "task-1251606" [ 1065.483024] env[61911]: _type = "Task" [ 1065.483024] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.490458] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251606, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.680460] env[61911]: DEBUG nova.compute.manager [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Received event network-changed-2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1065.680693] env[61911]: DEBUG nova.compute.manager [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Refreshing instance network info cache due to event network-changed-2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1065.680914] env[61911]: DEBUG oslo_concurrency.lockutils [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] Acquiring lock "refresh_cache-14986d4b-77af-461e-949e-fbc15f4ea6da" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.681222] env[61911]: DEBUG oslo_concurrency.lockutils [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] Acquired lock "refresh_cache-14986d4b-77af-461e-949e-fbc15f4ea6da" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.681417] env[61911]: DEBUG nova.network.neutron [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Refreshing network info cache for port 2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1065.774125] env[61911]: DEBUG nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1065.952657] env[61911]: DEBUG nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1065.993379] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251606, 'name': CreateVM_Task, 'duration_secs': 0.356018} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.993778] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1065.994324] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.994496] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.994807] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1065.995078] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f42fa32-eb14-474b-8964-c2ac6fcd3966 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.999849] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1065.999849] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525cfc9d-c929-b5ce-ee1d-d7ceb6e0a32d" [ 1065.999849] env[61911]: _type = "Task" [ 1065.999849] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.007495] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525cfc9d-c929-b5ce-ee1d-d7ceb6e0a32d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.293969] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.294259] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.296124] env[61911]: INFO nova.compute.claims [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.403389] env[61911]: DEBUG nova.network.neutron [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Updated VIF entry in instance network info cache for port 2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.403778] env[61911]: DEBUG nova.network.neutron [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Updating instance_info_cache with network_info: [{"id": "2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f", "address": "fa:16:3e:54:ee:05", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bbe32e9-4c", "ovs_interfaceid": "2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.471593] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.510077] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525cfc9d-c929-b5ce-ee1d-d7ceb6e0a32d, 'name': SearchDatastore_Task, 'duration_secs': 0.009179} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.510403] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.510644] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.510928] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.511141] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.511365] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.511679] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a738b1e-a8c7-4cee-8177-cd7ece215c4a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.519753] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1066.519929] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1066.520597] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4e9c37c-a806-4e76-8fda-3e1dd49be4d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.525460] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1066.525460] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524c7ed4-05f4-2c2e-f1b9-029a8ea78bf2" [ 1066.525460] env[61911]: _type = "Task" [ 1066.525460] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.534009] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524c7ed4-05f4-2c2e-f1b9-029a8ea78bf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.906881] env[61911]: DEBUG oslo_concurrency.lockutils [req-b662782b-f2c4-431c-b080-e93720be1055 req-cb3a7f94-8b67-48db-aedc-e50fa6d8e975 service nova] Releasing lock "refresh_cache-14986d4b-77af-461e-949e-fbc15f4ea6da" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.035515] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]524c7ed4-05f4-2c2e-f1b9-029a8ea78bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.008159} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.036342] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-479bf028-84db-4fbc-88b0-88ad6b7780fb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.041713] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1067.041713] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528699da-3401-e638-23fc-ebd2e8cc1771" [ 1067.041713] env[61911]: _type = "Task" [ 1067.041713] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.049014] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528699da-3401-e638-23fc-ebd2e8cc1771, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.437676] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0adf74-b0bd-4c50-a7e7-52ea6ffba06a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.445343] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07a81fc-7017-4fc3-b8b9-31ef4f9abadb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.474247] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a260fb03-b51e-4b4c-8ea2-ba51b5c3eea7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.482194] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c730249d-a852-409e-9b44-a502c888e03d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.495344] env[61911]: DEBUG nova.compute.provider_tree [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.552032] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528699da-3401-e638-23fc-ebd2e8cc1771, 'name': SearchDatastore_Task, 'duration_secs': 0.009165} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.552032] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.552032] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 14986d4b-77af-461e-949e-fbc15f4ea6da/14986d4b-77af-461e-949e-fbc15f4ea6da.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1067.552288] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfaa68bd-25f0-44d6-8019-d42560970809 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.558615] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1067.558615] env[61911]: value = "task-1251607" [ 1067.558615] env[61911]: _type = "Task" [ 1067.558615] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.567173] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251607, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.998233] env[61911]: DEBUG nova.scheduler.client.report [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1068.068112] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251607, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412327} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.068442] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 14986d4b-77af-461e-949e-fbc15f4ea6da/14986d4b-77af-461e-949e-fbc15f4ea6da.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1068.068595] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.068836] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-020b6987-8db0-4a96-89de-c777b01703ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.074761] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1068.074761] env[61911]: value = "task-1251608" [ 1068.074761] env[61911]: _type = "Task" [ 1068.074761] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.082328] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251608, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.503774] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.209s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.504344] env[61911]: DEBUG nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1068.507369] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.036s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.508821] env[61911]: INFO nova.compute.claims [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1068.584997] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251608, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063352} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.585346] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.586112] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be49c6e5-2fe1-45fd-9749-2bbee1653eb3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.607294] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 14986d4b-77af-461e-949e-fbc15f4ea6da/14986d4b-77af-461e-949e-fbc15f4ea6da.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.607809] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-962bd727-3eac-43fb-a9fb-63f27c17b6eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.627094] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1068.627094] env[61911]: value = "task-1251609" [ 1068.627094] env[61911]: _type = "Task" [ 1068.627094] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.634430] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251609, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.009162] env[61911]: DEBUG nova.compute.utils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1069.011057] env[61911]: DEBUG nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Not allocating networking since 'none' was specified. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1069.137243] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251609, 'name': ReconfigVM_Task, 'duration_secs': 0.331367} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.137538] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 14986d4b-77af-461e-949e-fbc15f4ea6da/14986d4b-77af-461e-949e-fbc15f4ea6da.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.138182] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9201ccc3-db7e-4c8b-abc7-5cc8b406e1cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.144441] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1069.144441] env[61911]: value = "task-1251610" [ 1069.144441] env[61911]: _type = "Task" [ 1069.144441] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.152022] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251610, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.513220] env[61911]: DEBUG nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1069.583134] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3dc4d8b-d825-4dd2-acf7-f4d2889fe688 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.590678] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb43a35-b3a4-4b0d-aee2-6fe40f077258 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.620791] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64555115-4cec-4c3f-b83f-9d5e026279b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.627571] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4bc3b9-6b1b-4cc1-8c96-5a44d25cc95a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.639904] env[61911]: DEBUG nova.compute.provider_tree [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.653272] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251610, 'name': Rename_Task, 'duration_secs': 0.136075} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.653517] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1069.653755] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-291aa87a-244c-4e07-86b1-12481dda8e35 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.659425] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1069.659425] env[61911]: value = "task-1251611" [ 1069.659425] env[61911]: _type = "Task" [ 1069.659425] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.666307] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251611, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.143021] env[61911]: DEBUG nova.scheduler.client.report [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1070.168236] env[61911]: DEBUG oslo_vmware.api [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251611, 'name': PowerOnVM_Task, 'duration_secs': 0.462091} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.168503] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1070.168703] env[61911]: INFO nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Took 6.68 seconds to spawn the instance on the hypervisor. [ 1070.168886] env[61911]: DEBUG nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1070.169643] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc0f99b-ba20-4b85-bf82-a940b43de966 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.525219] env[61911]: DEBUG nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1070.551599] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1070.551865] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.552043] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1070.552241] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.552398] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1070.552554] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1070.552800] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1070.552977] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1070.553173] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1070.553346] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1070.553529] env[61911]: DEBUG nova.virt.hardware [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1070.554454] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950e9814-62ab-4116-b482-5028b9010589 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.562434] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a855c9-b63e-4c7f-9817-371093391b9b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.576649] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.582036] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Creating folder: Project (7ed153d2b7cd4f66b4c1c35bfef7a85c). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1070.582306] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db38940b-b87f-4c2e-8c2a-de6954287f3b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.591620] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Created folder: Project (7ed153d2b7cd4f66b4c1c35bfef7a85c) in parent group-v269521. [ 1070.591801] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Creating folder: Instances. Parent ref: group-v269675. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1070.592020] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-60b9d313-c007-4392-962c-20e4d68c696b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.600129] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Created folder: Instances in parent group-v269675. [ 1070.600361] env[61911]: DEBUG oslo.service.loopingcall [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.600554] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1070.600744] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e9b81a3-e775-443c-84f9-40f3d20fe6f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.616031] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.616031] env[61911]: value = "task-1251614" [ 1070.616031] env[61911]: _type = "Task" [ 1070.616031] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.622898] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251614, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.647718] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.140s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.648270] env[61911]: DEBUG nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1070.684994] env[61911]: INFO nova.compute.manager [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Took 11.36 seconds to build instance. [ 1071.126249] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251614, 'name': CreateVM_Task, 'duration_secs': 0.348} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.126427] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.126858] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.127044] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.127367] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1071.127620] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beac6396-62e5-4d13-a18c-e4ea2b5081a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.131827] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1071.131827] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52616c73-591c-291d-afc5-0676e9ea7735" [ 1071.131827] env[61911]: _type = "Task" [ 1071.131827] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.139271] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52616c73-591c-291d-afc5-0676e9ea7735, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.152592] env[61911]: DEBUG nova.compute.utils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1071.153932] env[61911]: DEBUG nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Not allocating networking since 'none' was specified. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1071.186605] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a39374d0-75b2-4573-a570-2387c7c5b6f1 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.870s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.430827] env[61911]: DEBUG oslo_concurrency.lockutils [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "14986d4b-77af-461e-949e-fbc15f4ea6da" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.431137] env[61911]: DEBUG oslo_concurrency.lockutils [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.431331] env[61911]: DEBUG nova.compute.manager [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1071.432240] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01e4cc0-2742-4bc3-8ab9-7e053947e854 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.439301] env[61911]: DEBUG nova.compute.manager [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61911) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1071.439861] env[61911]: DEBUG nova.objects.instance [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'flavor' on Instance uuid 14986d4b-77af-461e-949e-fbc15f4ea6da {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.642800] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52616c73-591c-291d-afc5-0676e9ea7735, 'name': SearchDatastore_Task, 'duration_secs': 0.009158} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.642889] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.643114] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1071.643358] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.643512] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.643734] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.643957] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4dba027a-46b2-47e9-8b38-1abb96f6de4e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.651471] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.651661] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1071.652381] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10e92e7b-cfda-49e8-b907-b3ac6837299b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.654892] env[61911]: DEBUG nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1071.659871] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1071.659871] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527c97b2-a0cf-0bbd-3acb-516a2ba69767" [ 1071.659871] env[61911]: _type = "Task" [ 1071.659871] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.667206] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527c97b2-a0cf-0bbd-3acb-516a2ba69767, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.171017] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]527c97b2-a0cf-0bbd-3acb-516a2ba69767, 'name': SearchDatastore_Task, 'duration_secs': 0.010829} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.171797] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fe8b370-6178-4f70-a37c-7e6282e25064 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.176597] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1072.176597] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a9e973-6162-7645-dbcd-b5358ca5ec5e" [ 1072.176597] env[61911]: _type = "Task" [ 1072.176597] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.184029] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a9e973-6162-7645-dbcd-b5358ca5ec5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.446707] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.447084] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fd38632-d06e-4aae-a529-784f68b3eaaa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.454239] env[61911]: DEBUG oslo_vmware.api [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1072.454239] env[61911]: value = "task-1251615" [ 1072.454239] env[61911]: _type = "Task" [ 1072.454239] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.461592] env[61911]: DEBUG oslo_vmware.api [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.667386] env[61911]: DEBUG nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1072.686909] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a9e973-6162-7645-dbcd-b5358ca5ec5e, 'name': SearchDatastore_Task, 'duration_secs': 0.008637} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.688774] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.689058] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 1198724e-2802-47a0-a7e8-3941e3ffc175/1198724e-2802-47a0-a7e8-3941e3ffc175.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1072.689489] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-383f4ceb-d3ff-40e5-87cb-148b6f98aeca {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.695313] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1072.695548] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1072.695714] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1072.695906] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1072.696072] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1072.696232] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1072.696449] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1072.696614] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1072.696788] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1072.696987] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1072.697176] env[61911]: DEBUG nova.virt.hardware [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1072.697933] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fe3e7a-b652-46d8-9c99-344352457190 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.701687] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1072.701687] env[61911]: value = "task-1251616" [ 1072.701687] env[61911]: _type = "Task" [ 1072.701687] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.707865] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3f3941-9c0e-4f1b-8061-9a0ae671d7d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.714937] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251616, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.724580] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.730048] env[61911]: DEBUG oslo.service.loopingcall [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.730292] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1072.730511] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fe6f3b4-4a13-4fca-bf92-21236e9e567f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.746150] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.746150] env[61911]: value = "task-1251617" [ 1072.746150] env[61911]: _type = "Task" [ 1072.746150] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.754094] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251617, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.967728] env[61911]: DEBUG oslo_vmware.api [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251615, 'name': PowerOffVM_Task, 'duration_secs': 0.339898} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.968051] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1072.968266] env[61911]: DEBUG nova.compute.manager [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1072.969109] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03bf297-ba7e-4f79-b3a3-20c38b769d58 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.212595] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251616, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.255518] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251617, 'name': CreateVM_Task, 'duration_secs': 0.295725} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.255699] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1073.256145] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.256318] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.256641] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.256936] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82d904ce-3bdf-4c51-bad6-6d2cfd4c2223 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.261382] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1073.261382] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5275ec32-ac89-8b86-48c9-f1d1761050ca" [ 1073.261382] env[61911]: _type = "Task" [ 1073.261382] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.270252] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5275ec32-ac89-8b86-48c9-f1d1761050ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.481832] env[61911]: DEBUG oslo_concurrency.lockutils [None req-132814b0-6220-4fe9-a392-35af80e99794 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.051s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.712009] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251616, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539573} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.712324] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 1198724e-2802-47a0-a7e8-3941e3ffc175/1198724e-2802-47a0-a7e8-3941e3ffc175.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1073.712611] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.712872] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f0aa5f9-2f35-4810-84d4-5c7e9b884714 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.719130] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1073.719130] env[61911]: value = "task-1251618" [ 1073.719130] env[61911]: _type = "Task" [ 1073.719130] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.726569] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251618, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.769723] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5275ec32-ac89-8b86-48c9-f1d1761050ca, 'name': SearchDatastore_Task, 'duration_secs': 0.013364} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.770016] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.770268] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.770505] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.770658] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.770841] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.771103] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76c06b3a-e34c-444b-ad51-a709bb6e3182 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.778429] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.778606] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1073.779323] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74d84079-fd23-417d-9e25-6bbd33d7dc52 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.784176] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1073.784176] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52baa8f5-0362-22e7-355e-58c15772938c" [ 1073.784176] env[61911]: _type = "Task" [ 1073.784176] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.790943] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.791184] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.791390] env[61911]: INFO nova.compute.manager [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Shelving [ 1073.792764] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52baa8f5-0362-22e7-355e-58c15772938c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.071674] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "14986d4b-77af-461e-949e-fbc15f4ea6da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.071935] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.072167] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "14986d4b-77af-461e-949e-fbc15f4ea6da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.072359] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.072564] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.074786] env[61911]: INFO nova.compute.manager [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Terminating instance [ 1074.228866] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251618, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06761} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.229193] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1074.229892] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5668e4f-ec86-48fe-8143-843d628b893f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.248675] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 1198724e-2802-47a0-a7e8-3941e3ffc175/1198724e-2802-47a0-a7e8-3941e3ffc175.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.248903] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b272b780-ba13-4d96-92f2-5127354e1ecd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.266771] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1074.266771] env[61911]: value = "task-1251619" [ 1074.266771] env[61911]: _type = "Task" [ 1074.266771] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.275331] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251619, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.292658] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52baa8f5-0362-22e7-355e-58c15772938c, 'name': SearchDatastore_Task, 'duration_secs': 0.007796} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.295208] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ff57cc0-0dbe-4a6b-ba57-c7c43f62486a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.301437] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1074.301437] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52133472-a64f-7b9e-a27e-76fde6e0e976" [ 1074.301437] env[61911]: _type = "Task" [ 1074.301437] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.308733] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52133472-a64f-7b9e-a27e-76fde6e0e976, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.578088] env[61911]: DEBUG nova.compute.manager [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1074.578333] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1074.579237] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575231df-2130-47b0-8a6c-59a12481a104 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.586716] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.586943] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d712e58f-5b6e-4aaa-a991-4b018d0332ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.652201] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.652395] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.652620] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleting the datastore file [datastore2] 14986d4b-77af-461e-949e-fbc15f4ea6da {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.652903] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41c694b8-43a9-4f00-a765-b17f42e638db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.658689] env[61911]: DEBUG oslo_vmware.api [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1074.658689] env[61911]: value = "task-1251621" [ 1074.658689] env[61911]: _type = "Task" [ 1074.658689] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.666322] env[61911]: DEBUG oslo_vmware.api [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.776882] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251619, 'name': ReconfigVM_Task, 'duration_secs': 0.331172} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.777184] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 1198724e-2802-47a0-a7e8-3941e3ffc175/1198724e-2802-47a0-a7e8-3941e3ffc175.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.777842] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb8c946b-219d-46c9-b6d8-3b15c0d9bc03 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.783966] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1074.783966] env[61911]: value = "task-1251622" [ 1074.783966] env[61911]: _type = "Task" [ 1074.783966] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.791849] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251622, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.801506] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1074.801830] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34be5ead-df30-4686-ba88-1db7d7495234 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.811890] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52133472-a64f-7b9e-a27e-76fde6e0e976, 'name': SearchDatastore_Task, 'duration_secs': 0.01037} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.813133] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.813414] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.813764] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1074.813764] env[61911]: value = "task-1251623" [ 1074.813764] env[61911]: _type = "Task" [ 1074.813764] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.813966] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9cf71388-a044-4989-a043-e24aba47934a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.824835] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251623, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.826692] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1074.826692] env[61911]: value = "task-1251624" [ 1074.826692] env[61911]: _type = "Task" [ 1074.826692] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.833731] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251624, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.168958] env[61911]: DEBUG oslo_vmware.api [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150313} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.169275] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.169463] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.169647] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.169832] env[61911]: INFO nova.compute.manager [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1075.170102] env[61911]: DEBUG oslo.service.loopingcall [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.170315] env[61911]: DEBUG nova.compute.manager [-] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1075.170413] env[61911]: DEBUG nova.network.neutron [-] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1075.293759] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251622, 'name': Rename_Task, 'duration_secs': 0.166191} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.294090] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.294321] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ca005e4-96be-4e7f-9c45-57460ef2f030 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.299964] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1075.299964] env[61911]: value = "task-1251625" [ 1075.299964] env[61911]: _type = "Task" [ 1075.299964] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.309303] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.323831] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251623, 'name': PowerOffVM_Task, 'duration_secs': 0.178697} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.324139] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.324969] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77354e4a-579c-4de6-950f-2da9ba12c409 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.338383] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251624, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460092} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.352899] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1075.353225] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.353657] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66b1ada1-20c3-4656-b213-ce473557f9f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.356260] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6160c3ef-f4af-4a07-a40b-3d8d1b255346 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.367149] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1075.367149] env[61911]: value = "task-1251626" [ 1075.367149] env[61911]: _type = "Task" [ 1075.367149] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.375018] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.659966] env[61911]: DEBUG nova.compute.manager [req-12b68a4c-db55-4d73-9fea-dad05917a409 req-ab4c7304-02af-4778-8c36-d985ff55bf45 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Received event network-vif-deleted-2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1075.660191] env[61911]: INFO nova.compute.manager [req-12b68a4c-db55-4d73-9fea-dad05917a409 req-ab4c7304-02af-4778-8c36-d985ff55bf45 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Neutron deleted interface 2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f; detaching it from the instance and deleting it from the info cache [ 1075.660368] env[61911]: DEBUG nova.network.neutron [req-12b68a4c-db55-4d73-9fea-dad05917a409 req-ab4c7304-02af-4778-8c36-d985ff55bf45 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.809810] env[61911]: DEBUG oslo_vmware.api [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251625, 'name': PowerOnVM_Task, 'duration_secs': 0.44105} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.810138] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1075.810355] env[61911]: INFO nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Took 5.28 seconds to spawn the instance on the hypervisor. [ 1075.810542] env[61911]: DEBUG nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1075.811326] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5196a55a-e770-4586-a257-e0231b8efbb9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.868984] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1075.869608] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b6044b2f-d9aa-4eb9-bcdb-9f4abef14662 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.880364] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079116} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.881609] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1075.881964] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1075.881964] env[61911]: value = "task-1251627" [ 1075.881964] env[61911]: _type = "Task" [ 1075.881964] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.882688] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc8c792-76c3-4ceb-a52b-9b7c362154b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.904798] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1075.907781] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a87d26f4-b9ff-4008-939d-9fa3766df1a3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.921583] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251627, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.927041] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1075.927041] env[61911]: value = "task-1251628" [ 1075.927041] env[61911]: _type = "Task" [ 1075.927041] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.935868] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251628, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.135801] env[61911]: DEBUG nova.network.neutron [-] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.163712] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70a978ca-6383-4567-ad8a-eda7b2220395 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.172758] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008fd738-1f1a-4c29-a31e-f87ad6d4ae31 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.200065] env[61911]: DEBUG nova.compute.manager [req-12b68a4c-db55-4d73-9fea-dad05917a409 req-ab4c7304-02af-4778-8c36-d985ff55bf45 service nova] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Detach interface failed, port_id=2bbe32e9-4c4e-4428-8ed5-ef62d5531e4f, reason: Instance 14986d4b-77af-461e-949e-fbc15f4ea6da could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1076.326230] env[61911]: INFO nova.compute.manager [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Took 10.05 seconds to build instance. [ 1076.395106] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251627, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.435875] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251628, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.638580] env[61911]: INFO nova.compute.manager [-] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Took 1.47 seconds to deallocate network for instance. [ 1076.828610] env[61911]: DEBUG oslo_concurrency.lockutils [None req-31f33583-9903-415c-915e-c3b464b56f0f tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "1198724e-2802-47a0-a7e8-3941e3ffc175" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.557s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.896344] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251627, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.937631] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251628, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.145744] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.146208] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.146542] env[61911]: DEBUG nova.objects.instance [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'resources' on Instance uuid 14986d4b-77af-461e-949e-fbc15f4ea6da {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.397507] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251627, 'name': CreateSnapshot_Task, 'duration_secs': 1.303625} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.397870] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1077.398614] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcc8d55-7bf2-48fd-a7df-aeda677bc86e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.438367] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251628, 'name': ReconfigVM_Task, 'duration_secs': 1.263881} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.438672] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.439324] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-055991b5-1841-4ebb-b6c3-fdfed17ddd44 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.445080] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1077.445080] env[61911]: value = "task-1251629" [ 1077.445080] env[61911]: _type = "Task" [ 1077.445080] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.454740] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251629, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.736250] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799853a9-9ac1-4cfb-90fc-bbb039c61c88 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.743944] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42e51f6-2e73-4cb9-b7e1-2b8d379ce473 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.773280] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b87e1a-a4cc-492e-a0f6-63a3e5dd6839 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.780748] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03fe5746-3671-4e3e-9c94-c28d2a463cbb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.794279] env[61911]: DEBUG nova.compute.provider_tree [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.917327] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1077.917444] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8093985a-78d2-4cc5-b433-ed13f8484da2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.927563] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1077.927563] env[61911]: value = "task-1251630" [ 1077.927563] env[61911]: _type = "Task" [ 1077.927563] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.935089] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251630, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.955311] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251629, 'name': Rename_Task, 'duration_secs': 0.126289} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.955311] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1077.955311] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc652228-cc09-48c2-a723-4b90e675dad5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.960462] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1077.960462] env[61911]: value = "task-1251631" [ 1077.960462] env[61911]: _type = "Task" [ 1077.960462] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.967645] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.297207] env[61911]: DEBUG nova.scheduler.client.report [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.440207] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251630, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.469920] env[61911]: DEBUG oslo_vmware.api [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251631, 'name': PowerOnVM_Task, 'duration_secs': 0.493469} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.470166] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.470420] env[61911]: INFO nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Took 5.80 seconds to spawn the instance on the hypervisor. [ 1078.470635] env[61911]: DEBUG nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1078.471454] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed928d6-f788-4d50-8c05-9d1f85a713c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.803144] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.822200] env[61911]: INFO nova.scheduler.client.report [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted allocations for instance 14986d4b-77af-461e-949e-fbc15f4ea6da [ 1078.839053] env[61911]: DEBUG oslo_concurrency.lockutils [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.839322] env[61911]: DEBUG oslo_concurrency.lockutils [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.941422] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251630, 'name': CloneVM_Task} progress is 95%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.986189] env[61911]: INFO nova.compute.manager [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Took 12.53 seconds to build instance. [ 1079.329835] env[61911]: DEBUG oslo_concurrency.lockutils [None req-7bf85901-1b21-4e9e-b720-ddffa5965bc6 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "14986d4b-77af-461e-949e-fbc15f4ea6da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.258s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.342229] env[61911]: DEBUG nova.compute.utils [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.438422] env[61911]: INFO nova.compute.manager [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Rebuilding instance [ 1079.440653] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251630, 'name': CloneVM_Task} progress is 95%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.482403] env[61911]: DEBUG nova.compute.manager [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1079.483351] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0a88e9-1743-43b2-8852-930a675bb37e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.490298] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2f7e20f5-156c-4b61-8d5e-a73131d50c4c tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.039s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.844989] env[61911]: DEBUG oslo_concurrency.lockutils [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.938413] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251630, 'name': CloneVM_Task, 'duration_secs': 1.705317} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.938776] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Created linked-clone VM from snapshot [ 1079.939584] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33835c12-4b5c-41ed-9066-845f9e74848c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.946884] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Uploading image 7fc0c66b-5933-4158-bc37-1ff7cb26dee0 {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1079.969082] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1079.969082] env[61911]: value = "vm-269680" [ 1079.969082] env[61911]: _type = "VirtualMachine" [ 1079.969082] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1079.969682] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-47aac410-e4da-4c1a-96e1-ee4a7e2e8c51 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.976561] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease: (returnval){ [ 1079.976561] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5228a2a3-f813-7813-07f7-001eed85ec88" [ 1079.976561] env[61911]: _type = "HttpNfcLease" [ 1079.976561] env[61911]: } obtained for exporting VM: (result){ [ 1079.976561] env[61911]: value = "vm-269680" [ 1079.976561] env[61911]: _type = "VirtualMachine" [ 1079.976561] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1079.976797] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the lease: (returnval){ [ 1079.976797] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5228a2a3-f813-7813-07f7-001eed85ec88" [ 1079.976797] env[61911]: _type = "HttpNfcLease" [ 1079.976797] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1079.982731] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1079.982731] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5228a2a3-f813-7813-07f7-001eed85ec88" [ 1079.982731] env[61911]: _type = "HttpNfcLease" [ 1079.982731] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1080.071077] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "e183b035-9e04-44fe-99e2-3e943019a1f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.071327] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.485166] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1080.485166] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5228a2a3-f813-7813-07f7-001eed85ec88" [ 1080.485166] env[61911]: _type = "HttpNfcLease" [ 1080.485166] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1080.485672] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1080.485672] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5228a2a3-f813-7813-07f7-001eed85ec88" [ 1080.485672] env[61911]: _type = "HttpNfcLease" [ 1080.485672] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1080.486164] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c6de6e-e84f-463a-bc6d-a355905aa032 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.492883] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523bbc7f-08ab-33e8-e5b7-62fe15e12950/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1080.493068] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523bbc7f-08ab-33e8-e5b7-62fe15e12950/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1080.549563] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.550632] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e307ce8-a8c9-4217-be8d-85e9401ac316 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.558758] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1080.558758] env[61911]: value = "task-1251633" [ 1080.558758] env[61911]: _type = "Task" [ 1080.558758] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.566712] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.573232] env[61911]: DEBUG nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1080.586589] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-10d301b2-5887-487d-9d59-6e3b660f051e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.902058] env[61911]: DEBUG oslo_concurrency.lockutils [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.902288] env[61911]: DEBUG oslo_concurrency.lockutils [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.902560] env[61911]: INFO nova.compute.manager [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Attaching volume c0781785-8566-4c52-874a-48dcd75692cc to /dev/sdb [ 1080.933805] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9dadf4-c58f-480d-940f-b6dd1b32217d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.942473] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfd9dd4-c9a5-43c6-b00a-58b93b6c83b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.956173] env[61911]: DEBUG nova.virt.block_device [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Updating existing volume attachment record: 84f0fcc6-f9b2-4a98-9c9a-60c80b2d0d55 {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1081.068983] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251633, 'name': PowerOffVM_Task, 'duration_secs': 0.116929} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.069316] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.069625] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1081.070490] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0b8d3c-c8e9-4296-b553-3b78bd90cb6f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.080037] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.082142] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf47a74a-aaa1-47b9-b1c7-7e5026665f45 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.096792] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.097157] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.103179] env[61911]: INFO nova.compute.claims [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1081.109962] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.110369] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.111352] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleting the datastore file [datastore2] 7c2498d7-4e8c-4e81-a175-6bf8916b550e {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.111643] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b544f45d-983c-411d-8695-f90458059628 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.119399] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1081.119399] env[61911]: value = "task-1251635" [ 1081.119399] env[61911]: _type = "Task" [ 1081.119399] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.132136] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.629281] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251635, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102284} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.629655] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.629918] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.630312] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1082.184144] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11c6b35-f9af-4978-8ed7-48e389869319 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.192328] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531b4d88-dd73-45f3-af21-5c7fe4890729 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.224348] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458d6839-7795-491f-8bd5-5d3fd91bd7d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.232161] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c75da8-7426-4cf4-b670-f14d965c7ba1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.248078] env[61911]: DEBUG nova.compute.provider_tree [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.665106] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1082.665804] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.666116] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1082.668308] env[61911]: DEBUG nova.virt.hardware [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1082.668708] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c846814-704b-4054-8fc7-ae63a011f1c7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.677011] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91db7ceb-6f78-404c-9d74-d5d0c21b42b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.691048] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.696683] env[61911]: DEBUG oslo.service.loopingcall [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.696968] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1082.697210] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3bd87a3a-5b3f-4fcd-91b9-8811d69d487d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.713870] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.713870] env[61911]: value = "task-1251637" [ 1082.713870] env[61911]: _type = "Task" [ 1082.713870] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.721519] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251637, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.751720] env[61911]: DEBUG nova.scheduler.client.report [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1083.223848] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251637, 'name': CreateVM_Task, 'duration_secs': 0.343101} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.224050] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1083.224523] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.224709] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.225103] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.225428] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c025004c-5770-4882-9a98-29c5c250c594 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.230018] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1083.230018] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52844f46-2282-28f7-32a1-3942d6f8ec9d" [ 1083.230018] env[61911]: _type = "Task" [ 1083.230018] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.238111] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52844f46-2282-28f7-32a1-3942d6f8ec9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.257324] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.160s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.257995] env[61911]: DEBUG nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1083.740616] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52844f46-2282-28f7-32a1-3942d6f8ec9d, 'name': SearchDatastore_Task, 'duration_secs': 0.008994} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.740985] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.741222] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.741476] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.741626] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.741811] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.742105] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6b7212f-d0ef-4636-b94e-f7b133faa4c4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.750665] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.750870] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.751624] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ff7a644-3162-40ed-97d4-274d777d1a7b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.756670] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1083.756670] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525fb387-6cf7-1dff-c5c4-8432d8b1af60" [ 1083.756670] env[61911]: _type = "Task" [ 1083.756670] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.765734] env[61911]: DEBUG nova.compute.utils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1083.767085] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525fb387-6cf7-1dff-c5c4-8432d8b1af60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.767376] env[61911]: DEBUG nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1083.767566] env[61911]: DEBUG nova.network.neutron [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1083.814271] env[61911]: DEBUG nova.policy [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a5252f9309f4204a3a3fd7d865db6e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b285793da304b5e8f9736098ef163c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1084.060658] env[61911]: DEBUG nova.network.neutron [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Successfully created port: 269996dc-1a1b-416e-8543-4ea8147df4e3 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1084.270527] env[61911]: DEBUG nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1084.274189] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]525fb387-6cf7-1dff-c5c4-8432d8b1af60, 'name': SearchDatastore_Task, 'duration_secs': 0.01139} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.275566] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c04648c8-0ab2-4a36-9e8f-71f0cd338d00 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.282297] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1084.282297] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520e89b0-01fd-6071-27e6-d2952d098603" [ 1084.282297] env[61911]: _type = "Task" [ 1084.282297] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.294479] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520e89b0-01fd-6071-27e6-d2952d098603, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.792641] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]520e89b0-01fd-6071-27e6-d2952d098603, 'name': SearchDatastore_Task, 'duration_secs': 0.009392} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.792984] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.793220] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1084.793493] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f30fe47-e0e6-4cdf-b361-b6d1f45dbc97 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.799936] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1084.799936] env[61911]: value = "task-1251639" [ 1084.799936] env[61911]: _type = "Task" [ 1084.799936] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.808016] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251639, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.281631] env[61911]: DEBUG nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1085.307968] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1085.308242] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.308406] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1085.308596] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.308760] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1085.308918] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1085.309155] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1085.309320] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1085.309499] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1085.309667] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1085.309845] env[61911]: DEBUG nova.virt.hardware [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1085.310680] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0cfc6c-24c0-4445-a77a-d48d5031357d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.316318] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251639, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474259} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.316940] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.317193] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.317478] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-67a1264a-5d90-4b92-88bb-cf23ec206e38 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.322434] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a4e100-81a8-41b3-b931-875dc642df65 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.327407] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1085.327407] env[61911]: value = "task-1251640" [ 1085.327407] env[61911]: _type = "Task" [ 1085.327407] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.343723] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251640, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.435880] env[61911]: DEBUG nova.compute.manager [req-8f366652-3d7a-42d3-8a52-bc8309a11ebf req-c5546e92-3aac-44a2-8f47-e012d88019f9 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Received event network-vif-plugged-269996dc-1a1b-416e-8543-4ea8147df4e3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1085.436131] env[61911]: DEBUG oslo_concurrency.lockutils [req-8f366652-3d7a-42d3-8a52-bc8309a11ebf req-c5546e92-3aac-44a2-8f47-e012d88019f9 service nova] Acquiring lock "e183b035-9e04-44fe-99e2-3e943019a1f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.436352] env[61911]: DEBUG oslo_concurrency.lockutils [req-8f366652-3d7a-42d3-8a52-bc8309a11ebf req-c5546e92-3aac-44a2-8f47-e012d88019f9 service nova] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.436522] env[61911]: DEBUG oslo_concurrency.lockutils [req-8f366652-3d7a-42d3-8a52-bc8309a11ebf req-c5546e92-3aac-44a2-8f47-e012d88019f9 service nova] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.436727] env[61911]: DEBUG nova.compute.manager [req-8f366652-3d7a-42d3-8a52-bc8309a11ebf req-c5546e92-3aac-44a2-8f47-e012d88019f9 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] No waiting events found dispatching network-vif-plugged-269996dc-1a1b-416e-8543-4ea8147df4e3 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1085.436914] env[61911]: WARNING nova.compute.manager [req-8f366652-3d7a-42d3-8a52-bc8309a11ebf req-c5546e92-3aac-44a2-8f47-e012d88019f9 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Received unexpected event network-vif-plugged-269996dc-1a1b-416e-8543-4ea8147df4e3 for instance with vm_state building and task_state spawning. [ 1085.502536] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1085.502779] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269681', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'name': 'volume-c0781785-8566-4c52-874a-48dcd75692cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'serial': 'c0781785-8566-4c52-874a-48dcd75692cc'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1085.504051] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fdf536-94e3-443f-8349-37d943c54537 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.521659] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0cb8e7-526f-4141-adc5-d16c68eafd32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.546233] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] volume-c0781785-8566-4c52-874a-48dcd75692cc/volume-c0781785-8566-4c52-874a-48dcd75692cc.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.546561] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34ffcdb1-6197-4975-ae4e-32fcd99c9ba0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.565061] env[61911]: DEBUG oslo_vmware.api [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1085.565061] env[61911]: value = "task-1251641" [ 1085.565061] env[61911]: _type = "Task" [ 1085.565061] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.575944] env[61911]: DEBUG oslo_vmware.api [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251641, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.838767] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251640, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078176} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.839084] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.839849] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdff0323-04a1-4ee1-b350-edb59e72e427 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.860060] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.860468] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b706ef0-4584-4b08-ae3a-1f8a74f99cb9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.880479] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1085.880479] env[61911]: value = "task-1251642" [ 1085.880479] env[61911]: _type = "Task" [ 1085.880479] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.888819] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251642, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.011469] env[61911]: DEBUG nova.network.neutron [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Successfully updated port: 269996dc-1a1b-416e-8543-4ea8147df4e3 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1086.040967] env[61911]: DEBUG nova.compute.manager [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Received event network-changed-269996dc-1a1b-416e-8543-4ea8147df4e3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1086.041235] env[61911]: DEBUG nova.compute.manager [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Refreshing instance network info cache due to event network-changed-269996dc-1a1b-416e-8543-4ea8147df4e3. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1086.041473] env[61911]: DEBUG oslo_concurrency.lockutils [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] Acquiring lock "refresh_cache-e183b035-9e04-44fe-99e2-3e943019a1f9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.041621] env[61911]: DEBUG oslo_concurrency.lockutils [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] Acquired lock "refresh_cache-e183b035-9e04-44fe-99e2-3e943019a1f9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.041803] env[61911]: DEBUG nova.network.neutron [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Refreshing network info cache for port 269996dc-1a1b-416e-8543-4ea8147df4e3 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1086.075110] env[61911]: DEBUG oslo_vmware.api [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.389974] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251642, 'name': ReconfigVM_Task, 'duration_secs': 0.298751} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.390261] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 7c2498d7-4e8c-4e81-a175-6bf8916b550e/7c2498d7-4e8c-4e81-a175-6bf8916b550e.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.390841] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-22a1b914-7e45-4c73-82f3-68b3e9868244 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.397077] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1086.397077] env[61911]: value = "task-1251643" [ 1086.397077] env[61911]: _type = "Task" [ 1086.397077] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.408286] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251643, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.515600] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-e183b035-9e04-44fe-99e2-3e943019a1f9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.575312] env[61911]: DEBUG oslo_vmware.api [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251641, 'name': ReconfigVM_Task, 'duration_secs': 0.577426} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.576127] env[61911]: DEBUG nova.network.neutron [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1086.577858] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Reconfigured VM instance instance-00000064 to attach disk [datastore1] volume-c0781785-8566-4c52-874a-48dcd75692cc/volume-c0781785-8566-4c52-874a-48dcd75692cc.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.582481] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7c0b393-57f7-41c6-88ee-fa31d8462653 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.598901] env[61911]: DEBUG oslo_vmware.api [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1086.598901] env[61911]: value = "task-1251644" [ 1086.598901] env[61911]: _type = "Task" [ 1086.598901] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.608622] env[61911]: DEBUG oslo_vmware.api [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251644, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.736233] env[61911]: DEBUG nova.network.neutron [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.907951] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251643, 'name': Rename_Task, 'duration_secs': 0.140009} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.908273] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1086.908530] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-809084ec-1c80-4470-8d3c-1a5662c30b11 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.915292] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1086.915292] env[61911]: value = "task-1251645" [ 1086.915292] env[61911]: _type = "Task" [ 1086.915292] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.924178] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251645, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.110292] env[61911]: DEBUG oslo_vmware.api [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251644, 'name': ReconfigVM_Task, 'duration_secs': 0.164123} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.111029] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269681', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'name': 'volume-c0781785-8566-4c52-874a-48dcd75692cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'serial': 'c0781785-8566-4c52-874a-48dcd75692cc'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1087.239135] env[61911]: DEBUG oslo_concurrency.lockutils [req-71677e16-270c-46b5-ac35-f77872132280 req-14bf6abf-1805-4c2b-acee-265992611788 service nova] Releasing lock "refresh_cache-e183b035-9e04-44fe-99e2-3e943019a1f9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.239624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-e183b035-9e04-44fe-99e2-3e943019a1f9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.239818] env[61911]: DEBUG nova.network.neutron [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.426270] env[61911]: DEBUG oslo_vmware.api [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251645, 'name': PowerOnVM_Task, 'duration_secs': 0.505567} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.426560] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.426823] env[61911]: DEBUG nova.compute.manager [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1087.427713] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558eb6bb-dbfb-4552-8f0b-dc5d223733b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.771765] env[61911]: DEBUG nova.network.neutron [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1087.923576] env[61911]: DEBUG nova.network.neutron [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Updating instance_info_cache with network_info: [{"id": "269996dc-1a1b-416e-8543-4ea8147df4e3", "address": "fa:16:3e:9c:a2:d6", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269996dc-1a", "ovs_interfaceid": "269996dc-1a1b-416e-8543-4ea8147df4e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.944061] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.944398] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.944877] env[61911]: DEBUG nova.objects.instance [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1088.159661] env[61911]: DEBUG nova.objects.instance [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'flavor' on Instance uuid 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.426015] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.426488] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.426763] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.427011] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.427299] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.429456] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-e183b035-9e04-44fe-99e2-3e943019a1f9" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.429848] env[61911]: DEBUG nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Instance network_info: |[{"id": "269996dc-1a1b-416e-8543-4ea8147df4e3", "address": "fa:16:3e:9c:a2:d6", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269996dc-1a", "ovs_interfaceid": "269996dc-1a1b-416e-8543-4ea8147df4e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1088.430425] env[61911]: INFO nova.compute.manager [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Terminating instance [ 1088.432396] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:a2:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '269996dc-1a1b-416e-8543-4ea8147df4e3', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.440447] env[61911]: DEBUG oslo.service.loopingcall [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.442161] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1088.442772] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15b82595-264e-4863-9e29-9ae531fdfb80 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.467780] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.467780] env[61911]: value = "task-1251646" [ 1088.467780] env[61911]: _type = "Task" [ 1088.467780] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.477276] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251646, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.665821] env[61911]: DEBUG oslo_concurrency.lockutils [None req-914b4f58-0fbc-4f3f-8956-5632a7fb779e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.763s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.945311] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "refresh_cache-7c2498d7-4e8c-4e81-a175-6bf8916b550e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.945626] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "refresh_cache-7c2498d7-4e8c-4e81-a175-6bf8916b550e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.945680] env[61911]: DEBUG nova.network.neutron [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1088.964558] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3a4ff696-3164-433d-a03e-c8ecf42698e4 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.979650] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251646, 'name': CreateVM_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.123438] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523bbc7f-08ab-33e8-e5b7-62fe15e12950/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1089.124666] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f21a59e-c719-43f6-a6ef-3b63a22f83b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.131288] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523bbc7f-08ab-33e8-e5b7-62fe15e12950/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1089.131467] env[61911]: ERROR oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523bbc7f-08ab-33e8-e5b7-62fe15e12950/disk-0.vmdk due to incomplete transfer. [ 1089.131712] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e546df0d-9466-4401-b4d7-a0a58941f24f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.141848] env[61911]: DEBUG oslo_vmware.rw_handles [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/523bbc7f-08ab-33e8-e5b7-62fe15e12950/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1089.142091] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Uploaded image 7fc0c66b-5933-4158-bc37-1ff7cb26dee0 to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1089.144747] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1089.145037] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-20608722-2cde-43c9-b61c-21bcb0cc60b0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.151557] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1089.151557] env[61911]: value = "task-1251647" [ 1089.151557] env[61911]: _type = "Task" [ 1089.151557] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.159812] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251647, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.451825] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.452096] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.467659] env[61911]: DEBUG nova.network.neutron [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1089.479498] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251646, 'name': CreateVM_Task, 'duration_secs': 0.542396} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.481419] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1089.482095] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.482304] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.482674] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1089.483258] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abfc1ac4-babc-4735-91cf-02aee08dce87 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.487923] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1089.487923] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad0b94-e2e4-a0ec-e517-d7212de16e44" [ 1089.487923] env[61911]: _type = "Task" [ 1089.487923] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.495969] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad0b94-e2e4-a0ec-e517-d7212de16e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.525334] env[61911]: DEBUG nova.network.neutron [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.661209] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251647, 'name': Destroy_Task, 'duration_secs': 0.502255} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.661483] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Destroyed the VM [ 1089.661726] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1089.661972] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5ec58243-ca8e-4287-b3b8-5a85d47e8b38 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.667856] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1089.667856] env[61911]: value = "task-1251648" [ 1089.667856] env[61911]: _type = "Task" [ 1089.667856] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.675151] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251648, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.956371] env[61911]: INFO nova.compute.manager [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Detaching volume c0781785-8566-4c52-874a-48dcd75692cc [ 1089.988732] env[61911]: INFO nova.virt.block_device [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Attempting to driver detach volume c0781785-8566-4c52-874a-48dcd75692cc from mountpoint /dev/sdb [ 1089.989044] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1089.989274] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269681', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'name': 'volume-c0781785-8566-4c52-874a-48dcd75692cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'serial': 'c0781785-8566-4c52-874a-48dcd75692cc'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1089.990519] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51362f11-67be-4c8b-b7aa-8064a1ebac7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.002384] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ad0b94-e2e4-a0ec-e517-d7212de16e44, 'name': SearchDatastore_Task, 'duration_secs': 0.010539} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.018389] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.018665] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1090.018907] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.019081] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.019294] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1090.019681] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-395f52b0-0f9a-483c-95f2-fdc03b307f53 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.022212] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842ddf29-9c93-47a6-84ef-f06fb01246b5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.027977] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "refresh_cache-7c2498d7-4e8c-4e81-a175-6bf8916b550e" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.028420] env[61911]: DEBUG nova.compute.manager [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1090.028698] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1090.029546] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc981248-053d-46df-b763-da0ffff66fc7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.034979] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1090.035187] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1090.036710] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d3eb24-e8df-4b82-bfcb-36f475d02a91 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.039127] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cf77397-6e8a-4aab-b952-8c78dc6542a7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.044854] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.045626] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-712e9010-c140-42a4-b8f9-bb0bf6c1f19a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.066009] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abcb08e-d712-44fb-b148-0d794f7283f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.068834] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1090.068834] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52295cce-d0ba-3289-c3fa-a521da607492" [ 1090.068834] env[61911]: _type = "Task" [ 1090.068834] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.070237] env[61911]: DEBUG oslo_vmware.api [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1090.070237] env[61911]: value = "task-1251649" [ 1090.070237] env[61911]: _type = "Task" [ 1090.070237] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.086798] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] The volume has not been displaced from its original location: [datastore1] volume-c0781785-8566-4c52-874a-48dcd75692cc/volume-c0781785-8566-4c52-874a-48dcd75692cc.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1090.092212] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1090.092930] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-967d325b-9adc-480d-9145-2dabcc10a83c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.112218] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52295cce-d0ba-3289-c3fa-a521da607492, 'name': SearchDatastore_Task, 'duration_secs': 0.009894} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.112218] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdbac23c-c4d5-44e4-ba86-4b1ef37901f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.115803] env[61911]: DEBUG oslo_vmware.api [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.117261] env[61911]: DEBUG oslo_vmware.api [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1090.117261] env[61911]: value = "task-1251650" [ 1090.117261] env[61911]: _type = "Task" [ 1090.117261] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.122613] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1090.122613] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5207da55-9dff-e51d-c9ab-9fdfa696aea8" [ 1090.122613] env[61911]: _type = "Task" [ 1090.122613] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.125998] env[61911]: DEBUG oslo_vmware.api [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251650, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.132981] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5207da55-9dff-e51d-c9ab-9fdfa696aea8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.178082] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251648, 'name': RemoveSnapshot_Task, 'duration_secs': 0.386072} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.178417] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1090.178728] env[61911]: DEBUG nova.compute.manager [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1090.179574] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7f5c9d-3b14-423d-a6ad-f3270d6e0621 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.583726] env[61911]: DEBUG oslo_vmware.api [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251649, 'name': PowerOffVM_Task, 'duration_secs': 0.160347} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.584009] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1090.584202] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1090.584443] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9a7a72e-e648-4741-9cb0-fe237ca94e56 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.611283] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1090.611490] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1090.611675] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleting the datastore file [datastore1] 7c2498d7-4e8c-4e81-a175-6bf8916b550e {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.611917] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29feadb5-693c-4370-81b5-e69d619358f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.618357] env[61911]: DEBUG oslo_vmware.api [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1090.618357] env[61911]: value = "task-1251652" [ 1090.618357] env[61911]: _type = "Task" [ 1090.618357] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.628161] env[61911]: DEBUG oslo_vmware.api [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251650, 'name': ReconfigVM_Task, 'duration_secs': 0.20941} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.633509] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1090.638362] env[61911]: DEBUG oslo_vmware.api [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251652, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.638578] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75b21083-60a0-4cfc-aa99-06839641f3be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.652837] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5207da55-9dff-e51d-c9ab-9fdfa696aea8, 'name': SearchDatastore_Task, 'duration_secs': 0.009241} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.653971] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.654255] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] e183b035-9e04-44fe-99e2-3e943019a1f9/e183b035-9e04-44fe-99e2-3e943019a1f9.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1090.654578] env[61911]: DEBUG oslo_vmware.api [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1090.654578] env[61911]: value = "task-1251653" [ 1090.654578] env[61911]: _type = "Task" [ 1090.654578] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.654773] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5cee78c-48e3-49b1-b12a-c35858c8951a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.665696] env[61911]: DEBUG oslo_vmware.api [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251653, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.666804] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1090.666804] env[61911]: value = "task-1251654" [ 1090.666804] env[61911]: _type = "Task" [ 1090.666804] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.674133] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251654, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.691322] env[61911]: INFO nova.compute.manager [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Shelve offloading [ 1091.130372] env[61911]: DEBUG oslo_vmware.api [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251652, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112082} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.130631] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.130824] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1091.131017] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1091.131210] env[61911]: INFO nova.compute.manager [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1091.131454] env[61911]: DEBUG oslo.service.loopingcall [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.131648] env[61911]: DEBUG nova.compute.manager [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1091.131741] env[61911]: DEBUG nova.network.neutron [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1091.147864] env[61911]: DEBUG nova.network.neutron [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1091.166321] env[61911]: DEBUG oslo_vmware.api [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251653, 'name': ReconfigVM_Task, 'duration_secs': 0.146433} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.166631] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269681', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'name': 'volume-c0781785-8566-4c52-874a-48dcd75692cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0', 'attached_at': '', 'detached_at': '', 'volume_id': 'c0781785-8566-4c52-874a-48dcd75692cc', 'serial': 'c0781785-8566-4c52-874a-48dcd75692cc'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1091.176636] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251654, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.414911} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.176877] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore2] e183b035-9e04-44fe-99e2-3e943019a1f9/e183b035-9e04-44fe-99e2-3e943019a1f9.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1091.177106] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.177356] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54c83022-c4b5-4c36-b94e-75ad0c5283b8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.183507] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1091.183507] env[61911]: value = "task-1251655" [ 1091.183507] env[61911]: _type = "Task" [ 1091.183507] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.190609] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251655, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.195192] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.195522] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d479f073-2a87-4fee-b37a-ee8b987d42b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.200648] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1091.200648] env[61911]: value = "task-1251656" [ 1091.200648] env[61911]: _type = "Task" [ 1091.200648] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.208068] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251656, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.653069] env[61911]: DEBUG nova.network.neutron [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.693243] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251655, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062554} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.693745] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1091.694513] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2bcf0b4-950c-4e91-a19f-75430d664ea5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.716285] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] e183b035-9e04-44fe-99e2-3e943019a1f9/e183b035-9e04-44fe-99e2-3e943019a1f9.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1091.717623] env[61911]: DEBUG nova.objects.instance [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'flavor' on Instance uuid 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.721468] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cb227c9-dbcb-430e-aaf4-ff0c6349cbaa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.745417] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1091.745629] env[61911]: DEBUG nova.compute.manager [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1091.746524] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1091.746524] env[61911]: value = "task-1251657" [ 1091.746524] env[61911]: _type = "Task" [ 1091.746524] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.747214] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff83c743-1c7d-438b-a5a8-77588e50adbd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.756930] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.757071] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.757251] env[61911]: DEBUG nova.network.neutron [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1091.762924] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251657, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.155446] env[61911]: INFO nova.compute.manager [-] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Took 1.02 seconds to deallocate network for instance. [ 1092.240267] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b6ba7886-2704-4516-af9f-917c5b1fbb9e tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 2.788s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.260029] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251657, 'name': ReconfigVM_Task, 'duration_secs': 0.278671} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.260192] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Reconfigured VM instance instance-00000068 to attach disk [datastore2] e183b035-9e04-44fe-99e2-3e943019a1f9/e183b035-9e04-44fe-99e2-3e943019a1f9.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.260835] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1de33849-086e-4e84-a99e-3de754931a41 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.267125] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1092.267125] env[61911]: value = "task-1251658" [ 1092.267125] env[61911]: _type = "Task" [ 1092.267125] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.275447] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251658, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.519275] env[61911]: DEBUG nova.network.neutron [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19140385-69", "ovs_interfaceid": "19140385-696d-4e4c-a2dd-d53c936d4416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.664783] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.665075] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.665305] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.665499] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.665677] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.667933] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.668187] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.668409] env[61911]: DEBUG nova.objects.instance [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lazy-loading 'resources' on Instance uuid 7c2498d7-4e8c-4e81-a175-6bf8916b550e {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.669748] env[61911]: INFO nova.compute.manager [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Terminating instance [ 1092.777702] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251658, 'name': Rename_Task, 'duration_secs': 0.135615} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.777979] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1092.778249] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-620953b2-9735-4e56-b92e-ecea25777557 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.784528] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1092.784528] env[61911]: value = "task-1251659" [ 1092.784528] env[61911]: _type = "Task" [ 1092.784528] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.792172] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251659, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.022450] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.175971] env[61911]: DEBUG nova.compute.manager [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1093.175971] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.177267] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0eb12e-bb06-4666-934c-e614ebee75f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.185645] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1093.186056] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e02f936-12be-420c-b8ba-d46d90ffa601 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.193489] env[61911]: DEBUG oslo_vmware.api [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1093.193489] env[61911]: value = "task-1251660" [ 1093.193489] env[61911]: _type = "Task" [ 1093.193489] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.205826] env[61911]: DEBUG oslo_vmware.api [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.261350] env[61911]: DEBUG nova.compute.manager [req-50846eeb-8266-4616-b684-46c9b670aec9 req-53722f53-ea13-4469-baac-1ad6a254181d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-vif-unplugged-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1093.261636] env[61911]: DEBUG oslo_concurrency.lockutils [req-50846eeb-8266-4616-b684-46c9b670aec9 req-53722f53-ea13-4469-baac-1ad6a254181d service nova] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.262016] env[61911]: DEBUG oslo_concurrency.lockutils [req-50846eeb-8266-4616-b684-46c9b670aec9 req-53722f53-ea13-4469-baac-1ad6a254181d service nova] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.262180] env[61911]: DEBUG oslo_concurrency.lockutils [req-50846eeb-8266-4616-b684-46c9b670aec9 req-53722f53-ea13-4469-baac-1ad6a254181d service nova] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.262310] env[61911]: DEBUG nova.compute.manager [req-50846eeb-8266-4616-b684-46c9b670aec9 req-53722f53-ea13-4469-baac-1ad6a254181d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] No waiting events found dispatching network-vif-unplugged-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1093.262478] env[61911]: WARNING nova.compute.manager [req-50846eeb-8266-4616-b684-46c9b670aec9 req-53722f53-ea13-4469-baac-1ad6a254181d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received unexpected event network-vif-unplugged-19140385-696d-4e4c-a2dd-d53c936d4416 for instance with vm_state shelved and task_state shelving_offloading. [ 1093.266292] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317b21db-3dd1-452b-a434-81a5e7e22333 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.277659] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381ff0f8-b992-4a0b-8db1-aa3333311def {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.311136] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd422bb0-9546-4864-8523-542538e9b317 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.319249] env[61911]: DEBUG oslo_vmware.api [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251659, 'name': PowerOnVM_Task, 'duration_secs': 0.465041} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.322025] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1093.322025] env[61911]: INFO nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Took 8.04 seconds to spawn the instance on the hypervisor. [ 1093.322025] env[61911]: DEBUG nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1093.322687] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7b69f1-024d-42a7-abe9-01364c0fda2d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.326217] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41417d5a-0e45-4b6d-80e7-446f753d2f93 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.344475] env[61911]: DEBUG nova.compute.provider_tree [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.406836] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.407775] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f555193-29dd-4f69-a1c4-2c97b1f4cb2f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.416903] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.417247] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e12087cd-18b2-450f-80c7-db5393cc92d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.481398] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1093.481633] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1093.481810] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleting the datastore file [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.482119] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12c44c1e-dfda-4302-95af-2673b7067132 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.488668] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1093.488668] env[61911]: value = "task-1251662" [ 1093.488668] env[61911]: _type = "Task" [ 1093.488668] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.495782] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251662, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.705761] env[61911]: DEBUG oslo_vmware.api [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251660, 'name': PowerOffVM_Task, 'duration_secs': 0.183522} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.706060] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.706243] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.706488] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ec654dd-c410-449b-a386-60379adb02f0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.851283] env[61911]: DEBUG nova.scheduler.client.report [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1093.857754] env[61911]: INFO nova.compute.manager [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Took 12.78 seconds to build instance. [ 1093.998601] env[61911]: DEBUG oslo_vmware.api [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251662, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149656} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.998869] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.999069] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1093.999258] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1094.017519] env[61911]: INFO nova.scheduler.client.report [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted allocations for instance 95853730-1908-42f4-b30f-a4f276d2da8b [ 1094.356886] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.688s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.362483] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6cdcbc42-d463-491f-b5c2-a166618fe3f8 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.291s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.365407] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4819813b-b07a-4a57-b897-a920c1ad7cd1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.371830] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1dfc4a19-d649-451d-b21d-cba89b8a7d37 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Suspending the VM {{(pid=61911) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1094.372146] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ddb12d38-0efc-4076-a2d2-070fd5412cbf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.376607] env[61911]: INFO nova.scheduler.client.report [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleted allocations for instance 7c2498d7-4e8c-4e81-a175-6bf8916b550e [ 1094.378558] env[61911]: DEBUG oslo_vmware.api [None req-1dfc4a19-d649-451d-b21d-cba89b8a7d37 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1094.378558] env[61911]: value = "task-1251664" [ 1094.378558] env[61911]: _type = "Task" [ 1094.378558] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.388635] env[61911]: DEBUG oslo_vmware.api [None req-1dfc4a19-d649-451d-b21d-cba89b8a7d37 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251664, 'name': SuspendVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.521758] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.522143] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.522431] env[61911]: DEBUG nova.objects.instance [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'resources' on Instance uuid 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.708107] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1094.708107] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1094.708107] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleting the datastore file [datastore1] 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.708386] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08fb0ffa-bcf7-4e74-98ec-6c3e657d976b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.714809] env[61911]: DEBUG oslo_vmware.api [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1094.714809] env[61911]: value = "task-1251665" [ 1094.714809] env[61911]: _type = "Task" [ 1094.714809] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.722283] env[61911]: DEBUG oslo_vmware.api [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251665, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.890209] env[61911]: DEBUG oslo_vmware.api [None req-1dfc4a19-d649-451d-b21d-cba89b8a7d37 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251664, 'name': SuspendVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.890730] env[61911]: DEBUG oslo_concurrency.lockutils [None req-94877a03-cb65-44f4-88e0-1deb8d71e9f2 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "7c2498d7-4e8c-4e81-a175-6bf8916b550e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.464s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.025314] env[61911]: DEBUG nova.objects.instance [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'numa_topology' on Instance uuid 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.225116] env[61911]: DEBUG oslo_vmware.api [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251665, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212093} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.225432] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.225642] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1095.225826] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1095.226015] env[61911]: INFO nova.compute.manager [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Took 2.05 seconds to destroy the instance on the hypervisor. [ 1095.226282] env[61911]: DEBUG oslo.service.loopingcall [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.226497] env[61911]: DEBUG nova.compute.manager [-] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1095.226594] env[61911]: DEBUG nova.network.neutron [-] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1095.289583] env[61911]: DEBUG nova.compute.manager [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1095.289786] env[61911]: DEBUG nova.compute.manager [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing instance network info cache due to event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1095.290065] env[61911]: DEBUG oslo_concurrency.lockutils [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] Acquiring lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.290236] env[61911]: DEBUG oslo_concurrency.lockutils [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] Acquired lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.290408] env[61911]: DEBUG nova.network.neutron [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1095.394255] env[61911]: DEBUG oslo_vmware.api [None req-1dfc4a19-d649-451d-b21d-cba89b8a7d37 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251664, 'name': SuspendVM_Task, 'duration_secs': 0.554813} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.394736] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1dfc4a19-d649-451d-b21d-cba89b8a7d37 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Suspended the VM {{(pid=61911) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1095.394927] env[61911]: DEBUG nova.compute.manager [None req-1dfc4a19-d649-451d-b21d-cba89b8a7d37 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1095.395730] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853d4f8d-4db6-4948-ab98-12023ab6deea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.527758] env[61911]: DEBUG nova.objects.base [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Object Instance<95853730-1908-42f4-b30f-a4f276d2da8b> lazy-loaded attributes: resources,numa_topology {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1095.592194] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac5722a-8225-4976-9098-4ab4806ac9ce {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.599989] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa01667e-dda5-44e5-b1c7-f09955863073 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.630645] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd00d0e-cc1c-48c1-8729-039ecf364f77 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.638299] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d3fe66-20f6-4c43-b55c-f08977841b18 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.651022] env[61911]: DEBUG nova.compute.provider_tree [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.738066] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.738066] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.738066] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1095.987222] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "1198724e-2802-47a0-a7e8-3941e3ffc175" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.987499] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "1198724e-2802-47a0-a7e8-3941e3ffc175" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.987710] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "1198724e-2802-47a0-a7e8-3941e3ffc175-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.987893] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "1198724e-2802-47a0-a7e8-3941e3ffc175-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.988106] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "1198724e-2802-47a0-a7e8-3941e3ffc175-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.991458] env[61911]: INFO nova.compute.manager [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Terminating instance [ 1096.000697] env[61911]: DEBUG nova.network.neutron [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updated VIF entry in instance network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1096.001023] env[61911]: DEBUG nova.network.neutron [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap19140385-69", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.155041] env[61911]: DEBUG nova.scheduler.client.report [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.202460] env[61911]: DEBUG nova.network.neutron [-] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.241131] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.375551] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.495757] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "refresh_cache-1198724e-2802-47a0-a7e8-3941e3ffc175" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.496045] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquired lock "refresh_cache-1198724e-2802-47a0-a7e8-3941e3ffc175" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.496126] env[61911]: DEBUG nova.network.neutron [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1096.503414] env[61911]: DEBUG oslo_concurrency.lockutils [req-ebbc4ced-fcb9-40cd-afa9-73dcb69ce9bc req-c45696cb-2b2f-4f72-9b06-fe9c9ffd6e3d service nova] Releasing lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.659540] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.137s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.661953] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.421s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.662170] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.662382] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1096.664051] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f147dce-ffb6-42f8-9041-8dbe2bc188e4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.671705] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95881301-6458-47ae-b62b-983369d3ca4c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.685374] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e36580-d3f2-448c-ad46-071fe1f92b66 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.691469] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b448ea29-05f1-4841-8887-3202c7f01d6d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.721772] env[61911]: INFO nova.compute.manager [-] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Took 1.50 seconds to deallocate network for instance. [ 1096.722168] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180847MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1096.722322] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.722516] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.751342] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "e183b035-9e04-44fe-99e2-3e943019a1f9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.751560] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.751763] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "e183b035-9e04-44fe-99e2-3e943019a1f9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.751948] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.752230] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.754091] env[61911]: INFO nova.compute.manager [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Terminating instance [ 1097.014477] env[61911]: DEBUG nova.network.neutron [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1097.064964] env[61911]: DEBUG nova.network.neutron [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.167812] env[61911]: DEBUG oslo_concurrency.lockutils [None req-54d0ed10-591c-48a8-9791-116982cecd54 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.376s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.168605] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.793s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.168809] env[61911]: INFO nova.compute.manager [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Unshelving [ 1097.232185] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.257690] env[61911]: DEBUG nova.compute.manager [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1097.257899] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.258794] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9253f19-b819-4db1-8c67-7301c7326e51 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.266794] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.266993] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8dbd0e5-da1b-456f-812a-7939ba9c7016 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.319383] env[61911]: DEBUG nova.compute.manager [req-794885d9-5a16-4254-beb5-592a3586d4db req-a4f84e0d-825e-4e03-8ae4-26d776710128 service nova] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Received event network-vif-deleted-ebc99f07-acb2-44d1-9655-7359b307ecee {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1097.331524] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.331743] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.331927] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleting the datastore file [datastore2] e183b035-9e04-44fe-99e2-3e943019a1f9 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.332233] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59da5b60-9d6e-4f73-b4e0-6ea211c88e57 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.338207] env[61911]: DEBUG oslo_vmware.api [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1097.338207] env[61911]: value = "task-1251667" [ 1097.338207] env[61911]: _type = "Task" [ 1097.338207] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.345962] env[61911]: DEBUG oslo_vmware.api [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.568071] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Releasing lock "refresh_cache-1198724e-2802-47a0-a7e8-3941e3ffc175" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.568422] env[61911]: DEBUG nova.compute.manager [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1097.568628] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.569548] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5cfc4a-6ec7-46da-8bb1-66b9c80ac3e7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.577099] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.577338] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c031af18-6939-4aad-bf9c-2fb43ab3b97d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.583510] env[61911]: DEBUG oslo_vmware.api [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1097.583510] env[61911]: value = "task-1251668" [ 1097.583510] env[61911]: _type = "Task" [ 1097.583510] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.590771] env[61911]: DEBUG oslo_vmware.api [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.748579] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1097.748748] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 1198724e-2802-47a0-a7e8-3941e3ffc175 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1097.748876] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance e183b035-9e04-44fe-99e2-3e943019a1f9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1097.848065] env[61911]: DEBUG oslo_vmware.api [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137616} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.848308] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.848497] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1097.848676] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1097.848848] env[61911]: INFO nova.compute.manager [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1097.849099] env[61911]: DEBUG oslo.service.loopingcall [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.849299] env[61911]: DEBUG nova.compute.manager [-] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1097.849395] env[61911]: DEBUG nova.network.neutron [-] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.094946] env[61911]: DEBUG oslo_vmware.api [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251668, 'name': PowerOffVM_Task, 'duration_secs': 0.140447} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.095260] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.095441] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.095696] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93333cb3-a331-4cc9-913d-f03cdd9e1c35 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.116779] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.116955] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Deleting contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.117157] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleting the datastore file [datastore2] 1198724e-2802-47a0-a7e8-3941e3ffc175 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.117696] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7bf5b41-4ffa-4899-8821-ca188291e804 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.124486] env[61911]: DEBUG oslo_vmware.api [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for the task: (returnval){ [ 1098.124486] env[61911]: value = "task-1251670" [ 1098.124486] env[61911]: _type = "Task" [ 1098.124486] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.132085] env[61911]: DEBUG oslo_vmware.api [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251670, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.179576] env[61911]: DEBUG nova.compute.utils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1098.252423] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 95853730-1908-42f4-b30f-a4f276d2da8b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1098.252627] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1098.252780] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1098.307717] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1706ff72-6f9c-4245-91ce-035942e83c71 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.315185] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2939a51b-4944-4e96-abe5-48591d6bc9e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.344053] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f116fa0f-8d31-400a-a4b9-1a589d2df02a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.350823] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c4f385-7e3d-41e2-aece-42ade1deb9ac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.363008] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.544542] env[61911]: DEBUG nova.network.neutron [-] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.634648] env[61911]: DEBUG oslo_vmware.api [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Task: {'id': task-1251670, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086689} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.634922] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.635110] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Deleted contents of the VM from datastore datastore2 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.635298] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.635475] env[61911]: INFO nova.compute.manager [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1098.635719] env[61911]: DEBUG oslo.service.loopingcall [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.635910] env[61911]: DEBUG nova.compute.manager [-] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1098.636017] env[61911]: DEBUG nova.network.neutron [-] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.651069] env[61911]: DEBUG nova.network.neutron [-] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1098.682820] env[61911]: INFO nova.virt.block_device [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Booting with volume f409eb89-364e-4162-a493-56d73a45446b at /dev/sdb [ 1098.717318] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d54f887-c0de-4347-987b-1ef2417ff260 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.728591] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ab7007-3916-4ac3-9db5-92e087089d96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.751605] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eae4e7dd-f276-41cc-9f51-c5efb662796a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.759204] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dee8b3b-0f22-4f5f-b3d2-b50b42788f99 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.782028] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1875b6ce-f41b-4f49-b41a-715e58fd2d5b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.787471] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a692217-cc55-4858-a1f5-c5dea073ab1d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.800060] env[61911]: DEBUG nova.virt.block_device [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating existing volume attachment record: 483cd378-c150-4431-a4d6-0ded005e182d {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1098.865848] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1099.047110] env[61911]: INFO nova.compute.manager [-] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Took 1.20 seconds to deallocate network for instance. [ 1099.153394] env[61911]: DEBUG nova.network.neutron [-] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.345250] env[61911]: DEBUG nova.compute.manager [req-786c0120-5cae-47a5-aae6-3ea1ac5af83f req-23587fe8-c02a-44c3-811c-2208bedae861 service nova] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Received event network-vif-deleted-269996dc-1a1b-416e-8543-4ea8147df4e3 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1099.370438] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1099.370631] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.648s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.370900] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.139s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.371199] env[61911]: DEBUG nova.objects.instance [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'resources' on Instance uuid 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.372224] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1099.372702] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Cleaning up deleted instances with incomplete migration {{(pid=61911) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11646}} [ 1099.554674] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.655671] env[61911]: INFO nova.compute.manager [-] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Took 1.02 seconds to deallocate network for instance. [ 1099.877286] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1099.937012] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2572cae9-d2c0-4ca2-8824-8018a1a40c3e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.944765] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdbb1ff-513a-4c91-ae91-5b5772c41025 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.975531] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984ef5d8-1683-4c29-b614-f27fc98b3a79 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.982785] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3849e92c-1156-4e3a-ab60-fa731c1200ec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.995614] env[61911]: DEBUG nova.compute.provider_tree [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.162192] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.498874] env[61911]: DEBUG nova.scheduler.client.report [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1101.004009] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.006291] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.452s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.006532] env[61911]: DEBUG nova.objects.instance [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'resources' on Instance uuid e183b035-9e04-44fe-99e2-3e943019a1f9 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.022884] env[61911]: INFO nova.scheduler.client.report [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted allocations for instance 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0 [ 1101.374126] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.374126] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.374126] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1101.374126] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Rebuilding the list of instances to heal {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1101.530741] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1482ad0-f796-4fb7-9b83-2905f072cbf0 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.865s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.556574] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7aa7d4-861a-4a59-af45-ffaf6ce446e6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.564615] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fb070d-72d8-468e-91b1-e1ac249f4316 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.594750] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1c9278-abe6-4039-8684-727b10c442c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.601594] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c32fcf-6b82-4cea-b4a3-b71911098962 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.614129] env[61911]: DEBUG nova.compute.provider_tree [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.899938] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-1198724e-2802-47a0-a7e8-3941e3ffc175" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.900031] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-1198724e-2802-47a0-a7e8-3941e3ffc175" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.900186] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1101.900369] env[61911]: DEBUG nova.objects.instance [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lazy-loading 'info_cache' on Instance uuid 1198724e-2802-47a0-a7e8-3941e3ffc175 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.117523] env[61911]: DEBUG nova.scheduler.client.report [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1102.622361] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.625430] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.463s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.625430] env[61911]: DEBUG nova.objects.instance [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lazy-loading 'resources' on Instance uuid 1198724e-2802-47a0-a7e8-3941e3ffc175 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.722516] env[61911]: INFO nova.scheduler.client.report [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted allocations for instance e183b035-9e04-44fe-99e2-3e943019a1f9 [ 1102.920876] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1103.167787] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ce0c07-1726-496e-b2ec-4ee64fef0604 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.175324] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d7173a-692f-494d-a54c-ac35823d2376 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.205388] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af1f6f4-af0c-4926-a6b8-04e23d356448 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.212451] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d744c53d-fb02-43f6-aa6c-c1c387cf82fd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.225324] env[61911]: DEBUG nova.compute.provider_tree [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.232108] env[61911]: DEBUG oslo_concurrency.lockutils [None req-35a06e14-8a66-4032-9a91-99f9648dafa4 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "e183b035-9e04-44fe-99e2-3e943019a1f9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.480s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.474205] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.733484] env[61911]: DEBUG nova.scheduler.client.report [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1103.976665] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-1198724e-2802-47a0-a7e8-3941e3ffc175" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.976882] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1103.977110] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1103.977287] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1103.977556] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1103.977624] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1103.977702] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1103.977855] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1103.977980] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Cleaning up deleted instances {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11608}} [ 1104.134407] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.134650] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.238261] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.257973] env[61911]: INFO nova.scheduler.client.report [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Deleted allocations for instance 1198724e-2802-47a0-a7e8-3941e3ffc175 [ 1104.396298] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.396577] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.396792] env[61911]: DEBUG nova.objects.instance [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'pci_requests' on Instance uuid 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.492388] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] There are 54 instances to clean {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11617}} [ 1104.492662] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: e183b035-9e04-44fe-99e2-3e943019a1f9] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1104.637617] env[61911]: DEBUG nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1104.727824] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.727922] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.764888] env[61911]: DEBUG oslo_concurrency.lockutils [None req-1010e9fd-e6ba-486b-8511-945e3dbd0e13 tempest-ServerShowV247Test-1531314253 tempest-ServerShowV247Test-1531314253-project-member] Lock "1198724e-2802-47a0-a7e8-3941e3ffc175" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.777s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.900133] env[61911]: DEBUG nova.objects.instance [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'numa_topology' on Instance uuid 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.996778] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 7c2498d7-4e8c-4e81-a175-6bf8916b550e] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1105.158719] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.230598] env[61911]: DEBUG nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1105.402575] env[61911]: INFO nova.compute.claims [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1105.499916] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 14986d4b-77af-461e-949e-fbc15f4ea6da] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1105.752716] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.003063] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8e44d8fc-5fcb-4e96-93e8-84a67cc37ac0] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1106.455516] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70987a62-d7d2-4dfe-8224-2468af1b4509 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.463053] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1da7109-afe2-42f7-9e0c-3fe3d84dd87d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.491585] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90dfbf01-e82f-43d1-a430-b42475b31cd0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.498047] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8ecc68-47fa-4494-9c51-00f988a33368 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.510543] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: a03413fe-e74d-42e4-83ea-a9c19318526e] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1106.513433] env[61911]: DEBUG nova.compute.provider_tree [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.013773] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: f44c597d-ffb8-446b-b814-65733d077dbf] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1107.016204] env[61911]: DEBUG nova.scheduler.client.report [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1107.520578] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 632e811e-b253-47ed-9d1e-6b2538ff804b] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1107.522867] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.126s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.525164] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.367s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.526582] env[61911]: INFO nova.compute.claims [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1107.560185] env[61911]: INFO nova.network.neutron [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating port 19140385-696d-4e4c-a2dd-d53c936d4416 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1108.026424] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: b594227d-1647-4e43-8c90-962a9e765388] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1108.530622] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: f63fa31b-8840-4f95-9514-51f7d5393b25] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1108.582185] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ad230d-c8ad-40ae-892e-0f253ba631be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.590181] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1864644a-9525-4aad-aee8-98a30c673aab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.619100] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c88a56-d66b-418e-bbcb-9d9153866a04 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.626289] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bda2d5f-cbb1-42e2-9fe2-f5d82b49627c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.638907] env[61911]: DEBUG nova.compute.provider_tree [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.953244] env[61911]: DEBUG nova.compute.manager [req-d8825341-e61e-4122-b58a-c533af16e760 req-7ea94729-e8cc-4f4f-98d7-db462bfd6eac service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-vif-plugged-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1108.953474] env[61911]: DEBUG oslo_concurrency.lockutils [req-d8825341-e61e-4122-b58a-c533af16e760 req-7ea94729-e8cc-4f4f-98d7-db462bfd6eac service nova] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.953692] env[61911]: DEBUG oslo_concurrency.lockutils [req-d8825341-e61e-4122-b58a-c533af16e760 req-7ea94729-e8cc-4f4f-98d7-db462bfd6eac service nova] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.953865] env[61911]: DEBUG oslo_concurrency.lockutils [req-d8825341-e61e-4122-b58a-c533af16e760 req-7ea94729-e8cc-4f4f-98d7-db462bfd6eac service nova] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.954060] env[61911]: DEBUG nova.compute.manager [req-d8825341-e61e-4122-b58a-c533af16e760 req-7ea94729-e8cc-4f4f-98d7-db462bfd6eac service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] No waiting events found dispatching network-vif-plugged-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1108.954255] env[61911]: WARNING nova.compute.manager [req-d8825341-e61e-4122-b58a-c533af16e760 req-7ea94729-e8cc-4f4f-98d7-db462bfd6eac service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received unexpected event network-vif-plugged-19140385-696d-4e4c-a2dd-d53c936d4416 for instance with vm_state shelved_offloaded and task_state spawning. [ 1109.035037] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.035281] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.035437] env[61911]: DEBUG nova.network.neutron [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1109.036702] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 3831af9a-fe2a-49d6-9e38-fc78e2616461] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1109.144289] env[61911]: DEBUG nova.scheduler.client.report [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1109.541716] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: d9234a5f-4d26-4bdc-8f32-3120830a8abe] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1109.648988] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.124s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.649530] env[61911]: DEBUG nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1109.652136] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.900s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.653815] env[61911]: INFO nova.compute.claims [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1109.796314] env[61911]: DEBUG nova.network.neutron [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19140385-69", "ovs_interfaceid": "19140385-696d-4e4c-a2dd-d53c936d4416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.045402] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: b86d1c45-a81d-4ef0-aca9-8c8e7440f6c8] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1110.158080] env[61911]: DEBUG nova.compute.utils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1110.162294] env[61911]: DEBUG nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1110.162477] env[61911]: DEBUG nova.network.neutron [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1110.200293] env[61911]: DEBUG nova.policy [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62586445653844078feeeb731bd16f62', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e6ca33655884f16b4c0a6fcdb31ec12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1110.298657] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.325595] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='aab98f96ee9cba0ca0f345589bc18e27',container_format='bare',created_at=2024-10-10T16:03:07Z,direct_url=,disk_format='vmdk',id=7fc0c66b-5933-4158-bc37-1ff7cb26dee0,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-323112555-shelved',owner='ce8675694c2841a58e87bb7250f3435a',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-10T16:03:22Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1110.325878] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1110.326080] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1110.326318] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1110.326490] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1110.326648] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1110.326881] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1110.327049] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1110.327226] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1110.327397] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1110.327595] env[61911]: DEBUG nova.virt.hardware [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1110.328483] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f6d20f-f429-4ece-bfdd-3a7e7af72821 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.337274] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59aeeddb-4593-4fd7-99de-88c095f8d95b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.351011] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:3e:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a58387dd-f438-4913-af6a-fafb734cd881', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19140385-696d-4e4c-a2dd-d53c936d4416', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.358659] env[61911]: DEBUG oslo.service.loopingcall [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.358905] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1110.359132] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1fc91da-06bf-4676-83b5-dec4514c472b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.377345] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.377345] env[61911]: value = "task-1251675" [ 1110.377345] env[61911]: _type = "Task" [ 1110.377345] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.384370] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251675, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.495578] env[61911]: DEBUG nova.network.neutron [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Successfully created port: 56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1110.550255] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: b1d2b205-81d7-410b-901d-ccd608bc5bec] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1110.664549] env[61911]: DEBUG nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1110.725056] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726fb5c3-d230-4e65-9482-97d54a5a1adb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.733762] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea78365-3e8d-4d54-b95e-ecc21c740b87 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.762838] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39a60b5-b68e-4549-9b05-97b49d4f6e60 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.769687] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6002f7-c734-455d-8446-b6961020c722 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.782662] env[61911]: DEBUG nova.compute.provider_tree [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.887066] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251675, 'name': CreateVM_Task, 'duration_secs': 0.344048} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.887251] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1110.887927] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.888114] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.888535] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1110.888876] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51c9543b-2914-4d53-addb-dd970d5dd12b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.892957] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1110.892957] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52eb2f4d-ecab-a470-b45d-f6574a1bc696" [ 1110.892957] env[61911]: _type = "Task" [ 1110.892957] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.900254] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52eb2f4d-ecab-a470-b45d-f6574a1bc696, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.976040] env[61911]: DEBUG nova.compute.manager [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1110.976278] env[61911]: DEBUG nova.compute.manager [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing instance network info cache due to event network-changed-19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1110.976473] env[61911]: DEBUG oslo_concurrency.lockutils [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] Acquiring lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.976881] env[61911]: DEBUG oslo_concurrency.lockutils [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] Acquired lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.976881] env[61911]: DEBUG nova.network.neutron [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Refreshing network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1111.052103] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 7060c2a1-7546-41f9-891d-af6a59fa1aa5] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1111.286046] env[61911]: DEBUG nova.scheduler.client.report [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1111.403683] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.403945] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Processing image 7fc0c66b-5933-4158-bc37-1ff7cb26dee0 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1111.404221] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0/7fc0c66b-5933-4158-bc37-1ff7cb26dee0.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.404378] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0/7fc0c66b-5933-4158-bc37-1ff7cb26dee0.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.404561] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1111.404808] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff28275e-b5cf-4b42-b88f-ef806e32e599 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.413853] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1111.414113] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1111.414813] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48bb39e9-7c8d-488a-8eca-8bbc284afd85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.419941] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1111.419941] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e340ff-d529-0371-9b09-51a421168980" [ 1111.419941] env[61911]: _type = "Task" [ 1111.419941] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.427864] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e340ff-d529-0371-9b09-51a421168980, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.555686] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: fddba36d-1b15-43fb-9e99-68880e8a235c] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1111.676798] env[61911]: DEBUG nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1111.685464] env[61911]: DEBUG nova.network.neutron [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updated VIF entry in instance network info cache for port 19140385-696d-4e4c-a2dd-d53c936d4416. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1111.686261] env[61911]: DEBUG nova.network.neutron [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [{"id": "19140385-696d-4e4c-a2dd-d53c936d4416", "address": "fa:16:3e:7c:3e:3f", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19140385-69", "ovs_interfaceid": "19140385-696d-4e4c-a2dd-d53c936d4416", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.704293] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1111.704548] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1111.704711] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1111.704896] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1111.705067] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1111.705228] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1111.705513] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1111.705785] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1111.706172] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1111.706267] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1111.706415] env[61911]: DEBUG nova.virt.hardware [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1111.707586] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1886bd1-2305-42f8-b78f-a59305021644 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.715525] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffdcfa2-9128-4703-9e1f-489e4f671fd0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.791053] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.139s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.791600] env[61911]: DEBUG nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1111.929599] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Preparing fetch location {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1111.930151] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Fetch image to [datastore1] OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507/OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507.vmdk {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1111.930377] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Downloading stream optimized image 7fc0c66b-5933-4158-bc37-1ff7cb26dee0 to [datastore1] OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507/OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507.vmdk on the data store datastore1 as vApp {{(pid=61911) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1111.930565] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Downloading image file data 7fc0c66b-5933-4158-bc37-1ff7cb26dee0 to the ESX as VM named 'OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507' {{(pid=61911) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1111.973830] env[61911]: DEBUG nova.network.neutron [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Successfully updated port: 56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1112.004770] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1112.004770] env[61911]: value = "resgroup-9" [ 1112.004770] env[61911]: _type = "ResourcePool" [ 1112.004770] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1112.005057] env[61911]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-241688b8-6db5-4b6d-bf75-aa88b77e8cb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.027761] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease: (returnval){ [ 1112.027761] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52315862-df08-87e6-6975-7cd2694bbf40" [ 1112.027761] env[61911]: _type = "HttpNfcLease" [ 1112.027761] env[61911]: } obtained for vApp import into resource pool (val){ [ 1112.027761] env[61911]: value = "resgroup-9" [ 1112.027761] env[61911]: _type = "ResourcePool" [ 1112.027761] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1112.027761] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the lease: (returnval){ [ 1112.027761] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52315862-df08-87e6-6975-7cd2694bbf40" [ 1112.027761] env[61911]: _type = "HttpNfcLease" [ 1112.027761] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1112.032550] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1112.032550] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52315862-df08-87e6-6975-7cd2694bbf40" [ 1112.032550] env[61911]: _type = "HttpNfcLease" [ 1112.032550] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1112.059712] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: ec53a8ee-df42-4003-a933-17aa4c90c3aa] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1112.189477] env[61911]: DEBUG oslo_concurrency.lockutils [req-3076b719-46b1-4e58-8189-9e8bb9fd9a9d req-b01fc2eb-3ce2-4313-968f-33d9a242e178 service nova] Releasing lock "refresh_cache-95853730-1908-42f4-b30f-a4f276d2da8b" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.296906] env[61911]: DEBUG nova.compute.utils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1112.298454] env[61911]: DEBUG nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1112.298643] env[61911]: DEBUG nova.network.neutron [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1112.338492] env[61911]: DEBUG nova.policy [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a5252f9309f4204a3a3fd7d865db6e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b285793da304b5e8f9736098ef163c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1112.476802] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.476802] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.477103] env[61911]: DEBUG nova.network.neutron [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1112.534888] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1112.534888] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52315862-df08-87e6-6975-7cd2694bbf40" [ 1112.534888] env[61911]: _type = "HttpNfcLease" [ 1112.534888] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1112.535215] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1112.535215] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52315862-df08-87e6-6975-7cd2694bbf40" [ 1112.535215] env[61911]: _type = "HttpNfcLease" [ 1112.535215] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1112.535925] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6516a30-1908-4d7d-bb85-3ff7cfff203a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.543084] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ebb9b-47f9-edc1-587b-a5fc4bbba62f/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1112.543275] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ebb9b-47f9-edc1-587b-a5fc4bbba62f/disk-0.vmdk. {{(pid=61911) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1112.603110] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: f386424d-953d-431d-a56b-8542f2a3458c] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1112.606162] env[61911]: DEBUG nova.network.neutron [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Successfully created port: 59c51f38-526f-4792-88c9-ba9bb0d19458 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1112.613357] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7077d147-f96c-4a78-8b0d-3e9afad155eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.802526] env[61911]: DEBUG nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1113.004126] env[61911]: DEBUG nova.compute.manager [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Received event network-vif-plugged-56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1113.004392] env[61911]: DEBUG oslo_concurrency.lockutils [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.004768] env[61911]: DEBUG oslo_concurrency.lockutils [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.004925] env[61911]: DEBUG oslo_concurrency.lockutils [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.005070] env[61911]: DEBUG nova.compute.manager [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] No waiting events found dispatching network-vif-plugged-56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1113.005260] env[61911]: WARNING nova.compute.manager [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Received unexpected event network-vif-plugged-56626d17-a6ef-4464-b86f-3ca4a0b20e09 for instance with vm_state building and task_state spawning. [ 1113.005424] env[61911]: DEBUG nova.compute.manager [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Received event network-changed-56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1113.005611] env[61911]: DEBUG nova.compute.manager [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Refreshing instance network info cache due to event network-changed-56626d17-a6ef-4464-b86f-3ca4a0b20e09. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1113.005822] env[61911]: DEBUG oslo_concurrency.lockutils [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] Acquiring lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.010360] env[61911]: DEBUG nova.network.neutron [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1113.106024] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 3c4c8943-5324-4c3e-b7e0-7ea5972bc026] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1113.158581] env[61911]: DEBUG nova.network.neutron [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updating instance_info_cache with network_info: [{"id": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "address": "fa:16:3e:45:8f:e1", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56626d17-a6", "ovs_interfaceid": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.610704] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8d9881fe-4bb5-4d85-9be9-32c2eca32a9a] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1113.661194] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.661592] env[61911]: DEBUG nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Instance network_info: |[{"id": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "address": "fa:16:3e:45:8f:e1", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56626d17-a6", "ovs_interfaceid": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1113.663393] env[61911]: DEBUG oslo_concurrency.lockutils [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] Acquired lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.663737] env[61911]: DEBUG nova.network.neutron [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Refreshing network info cache for port 56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1113.665185] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:8f:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56626d17-a6ef-4464-b86f-3ca4a0b20e09', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1113.673790] env[61911]: DEBUG oslo.service.loopingcall [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.678096] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1113.678392] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f2e649f-f019-4acb-9cdb-5bef07f44cf5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.704695] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1113.704695] env[61911]: value = "task-1251677" [ 1113.704695] env[61911]: _type = "Task" [ 1113.704695] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.713726] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251677, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.729134] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Completed reading data from the image iterator. {{(pid=61911) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1113.729400] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ebb9b-47f9-edc1-587b-a5fc4bbba62f/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1113.730589] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0605bd54-a7c5-455a-bb56-f5a2c18044d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.736683] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ebb9b-47f9-edc1-587b-a5fc4bbba62f/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1113.736871] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ebb9b-47f9-edc1-587b-a5fc4bbba62f/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1113.737167] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4f115a72-42a4-48f3-80cf-3f110f05ec4a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.815111] env[61911]: DEBUG nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1113.841348] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1113.841659] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.841843] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1113.842075] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.842260] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1113.842442] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1113.842680] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1113.842917] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1113.843140] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1113.843347] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1113.843545] env[61911]: DEBUG nova.virt.hardware [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1113.844568] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a41b297-3c77-4caf-9440-9d84aa4c9df8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.853548] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966a0013-de73-415a-b6bb-8bbedba34117 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.970114] env[61911]: DEBUG oslo_vmware.rw_handles [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ebb9b-47f9-edc1-587b-a5fc4bbba62f/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1113.970378] env[61911]: INFO nova.virt.vmwareapi.images [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Downloaded image file data 7fc0c66b-5933-4158-bc37-1ff7cb26dee0 [ 1113.971270] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf8ddc6-9ccf-48ec-85a8-889ce1ee716d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.987669] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2801cd33-1842-48c8-9c4b-bede9138cb01 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.009159] env[61911]: INFO nova.virt.vmwareapi.images [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] The imported VM was unregistered [ 1114.011947] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Caching image {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1114.012249] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating directory with path [datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.012662] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb77515b-8e35-4e80-97af-922c93e74186 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.023302] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created directory with path [datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.023456] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507/OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507.vmdk to [datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0/7fc0c66b-5933-4158-bc37-1ff7cb26dee0.vmdk. {{(pid=61911) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1114.023804] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e4576ac9-b5f7-4797-9086-1f6fe89fb1bf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.029574] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1114.029574] env[61911]: value = "task-1251679" [ 1114.029574] env[61911]: _type = "Task" [ 1114.029574] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.037278] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251679, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.114427] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 0ebf8d34-54c0-453e-b2e2-820feed6ee61] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1114.161521] env[61911]: DEBUG nova.network.neutron [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Successfully updated port: 59c51f38-526f-4792-88c9-ba9bb0d19458 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1114.214057] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251677, 'name': CreateVM_Task, 'duration_secs': 0.32066} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.214057] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1114.214057] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.214268] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.214796] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1114.214796] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8045108c-1864-417a-999d-99ab1302d142 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.218747] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1114.218747] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526bd9ad-36f4-1e4d-88a9-f7cf227f61d0" [ 1114.218747] env[61911]: _type = "Task" [ 1114.218747] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.225810] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526bd9ad-36f4-1e4d-88a9-f7cf227f61d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.516170] env[61911]: DEBUG nova.network.neutron [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updated VIF entry in instance network info cache for port 56626d17-a6ef-4464-b86f-3ca4a0b20e09. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1114.516622] env[61911]: DEBUG nova.network.neutron [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updating instance_info_cache with network_info: [{"id": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "address": "fa:16:3e:45:8f:e1", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56626d17-a6", "ovs_interfaceid": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.540496] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251679, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.617637] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 75577b68-b012-43d5-abdb-bd5f54c249d5] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1114.664174] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.664361] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.664476] env[61911]: DEBUG nova.network.neutron [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1114.728999] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526bd9ad-36f4-1e4d-88a9-f7cf227f61d0, 'name': SearchDatastore_Task, 'duration_secs': 0.014589} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.729419] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.729766] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1114.730090] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.730375] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.730598] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.730878] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-410478e0-98e2-4be5-894a-9145711cfd18 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.746171] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.746407] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1114.747237] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b5d7792-8380-44fb-b0b1-7621917facbc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.752684] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1114.752684] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d11524-b145-7939-6520-3f13f18352b1" [ 1114.752684] env[61911]: _type = "Task" [ 1114.752684] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.760755] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d11524-b145-7939-6520-3f13f18352b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.019782] env[61911]: DEBUG oslo_concurrency.lockutils [req-08ed3776-9af7-47f4-bb68-f990006412bc req-6149a098-30bc-408b-9327-ec5c4664760d service nova] Releasing lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.035715] env[61911]: DEBUG nova.compute.manager [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Received event network-vif-plugged-59c51f38-526f-4792-88c9-ba9bb0d19458 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1115.035955] env[61911]: DEBUG oslo_concurrency.lockutils [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] Acquiring lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.036188] env[61911]: DEBUG oslo_concurrency.lockutils [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.036366] env[61911]: DEBUG oslo_concurrency.lockutils [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.036539] env[61911]: DEBUG nova.compute.manager [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] No waiting events found dispatching network-vif-plugged-59c51f38-526f-4792-88c9-ba9bb0d19458 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1115.036710] env[61911]: WARNING nova.compute.manager [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Received unexpected event network-vif-plugged-59c51f38-526f-4792-88c9-ba9bb0d19458 for instance with vm_state building and task_state spawning. [ 1115.036916] env[61911]: DEBUG nova.compute.manager [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Received event network-changed-59c51f38-526f-4792-88c9-ba9bb0d19458 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1115.037113] env[61911]: DEBUG nova.compute.manager [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Refreshing instance network info cache due to event network-changed-59c51f38-526f-4792-88c9-ba9bb0d19458. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1115.037291] env[61911]: DEBUG oslo_concurrency.lockutils [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] Acquiring lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.044245] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251679, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.121121] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: b8cd7073-c712-4470-8ece-d144e8ce275f] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1115.202338] env[61911]: DEBUG nova.network.neutron [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1115.263878] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d11524-b145-7939-6520-3f13f18352b1, 'name': SearchDatastore_Task, 'duration_secs': 0.077723} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.264731] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08af8b97-e020-4c7f-94fb-0a7f41802cc0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.270675] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1115.270675] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a8c0c0-72ef-598b-39ab-325004fd0e57" [ 1115.270675] env[61911]: _type = "Task" [ 1115.270675] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.282716] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a8c0c0-72ef-598b-39ab-325004fd0e57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.347525] env[61911]: DEBUG nova.network.neutron [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance_info_cache with network_info: [{"id": "59c51f38-526f-4792-88c9-ba9bb0d19458", "address": "fa:16:3e:c6:e2:6d", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59c51f38-52", "ovs_interfaceid": "59c51f38-526f-4792-88c9-ba9bb0d19458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.541589] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251679, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.624716] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 50a522ab-651a-4bb3-93d3-8866169e95e4] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1115.783537] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a8c0c0-72ef-598b-39ab-325004fd0e57, 'name': SearchDatastore_Task, 'duration_secs': 0.090098} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.783823] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.784175] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8ac6fded-77cd-4258-8fcf-c36f27435c20/8ac6fded-77cd-4258-8fcf-c36f27435c20.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1115.784517] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f604c307-f4a4-4ef1-91b0-cb5a3b2837a6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.791137] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1115.791137] env[61911]: value = "task-1251680" [ 1115.791137] env[61911]: _type = "Task" [ 1115.791137] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.798894] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251680, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.850721] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.851172] env[61911]: DEBUG nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Instance network_info: |[{"id": "59c51f38-526f-4792-88c9-ba9bb0d19458", "address": "fa:16:3e:c6:e2:6d", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59c51f38-52", "ovs_interfaceid": "59c51f38-526f-4792-88c9-ba9bb0d19458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1115.851573] env[61911]: DEBUG oslo_concurrency.lockutils [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] Acquired lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.851873] env[61911]: DEBUG nova.network.neutron [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Refreshing network info cache for port 59c51f38-526f-4792-88c9-ba9bb0d19458 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1115.853249] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:e2:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19440099-773e-4a31-b82e-84a4daa5d8fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59c51f38-526f-4792-88c9-ba9bb0d19458', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.860727] env[61911]: DEBUG oslo.service.loopingcall [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.863954] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1115.864821] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-187cc796-858c-427c-bca2-1e0647214b06 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.884777] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.884777] env[61911]: value = "task-1251681" [ 1115.884777] env[61911]: _type = "Task" [ 1115.884777] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.893399] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251681, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.041090] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251679, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.127890] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: a35a11f1-3d4b-439e-9517-fa5a9f43af60] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1116.151714] env[61911]: DEBUG nova.network.neutron [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updated VIF entry in instance network info cache for port 59c51f38-526f-4792-88c9-ba9bb0d19458. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1116.152162] env[61911]: DEBUG nova.network.neutron [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance_info_cache with network_info: [{"id": "59c51f38-526f-4792-88c9-ba9bb0d19458", "address": "fa:16:3e:c6:e2:6d", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59c51f38-52", "ovs_interfaceid": "59c51f38-526f-4792-88c9-ba9bb0d19458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.301281] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251680, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.394528] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251681, 'name': CreateVM_Task, 'duration_secs': 0.353085} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.394739] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1116.395646] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.395817] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.396184] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1116.396441] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c59420b-1b29-4919-8396-70f10eb79d5e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.400861] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1116.400861] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5227a589-9a89-1c05-bb2f-1a5903510126" [ 1116.400861] env[61911]: _type = "Task" [ 1116.400861] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.407994] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5227a589-9a89-1c05-bb2f-1a5903510126, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.543371] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251679, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.347048} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.543672] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507/OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507.vmdk to [datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0/7fc0c66b-5933-4158-bc37-1ff7cb26dee0.vmdk. [ 1116.543871] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Cleaning up location [datastore1] OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1116.544053] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_3cf2ffd1-f911-4dd9-b29a-658fc437e507 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.544326] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73dd7c36-4b6f-4c22-96b4-4cb7374a3f09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.551016] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1116.551016] env[61911]: value = "task-1251682" [ 1116.551016] env[61911]: _type = "Task" [ 1116.551016] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.558859] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251682, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.632116] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 5997c8e4-61d4-4043-99ab-5e1c5e5dcc7c] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1116.655419] env[61911]: DEBUG oslo_concurrency.lockutils [req-c44e8719-e199-4139-a486-8f4a7cb7de03 req-f9cbec8d-28e3-4d81-8a26-0bea434bff5f service nova] Releasing lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.801650] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251680, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.986083} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.801950] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 8ac6fded-77cd-4258-8fcf-c36f27435c20/8ac6fded-77cd-4258-8fcf-c36f27435c20.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1116.802202] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1116.802514] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ad6df71-c1f5-46ca-ab10-f7514bf3e6d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.808803] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1116.808803] env[61911]: value = "task-1251683" [ 1116.808803] env[61911]: _type = "Task" [ 1116.808803] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.816695] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.911470] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5227a589-9a89-1c05-bb2f-1a5903510126, 'name': SearchDatastore_Task, 'duration_secs': 0.011533} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.911627] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.911769] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1116.912028] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.912189] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.912371] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1116.912624] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7b256d6-9fc7-4b00-95c0-af53b2f54749 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.920428] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1116.920601] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1116.921276] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e7ce534-6a5e-41e8-9462-a79333357ae1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.926138] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1116.926138] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5248dd71-3002-3ac5-1c74-7c3680a26113" [ 1116.926138] env[61911]: _type = "Task" [ 1116.926138] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.933128] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5248dd71-3002-3ac5-1c74-7c3680a26113, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.060237] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251682, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190302} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.060436] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.060609] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0/7fc0c66b-5933-4158-bc37-1ff7cb26dee0.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.060854] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0/7fc0c66b-5933-4158-bc37-1ff7cb26dee0.vmdk to [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1117.061115] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-141d7edc-af71-4e9a-bfde-87b389e6e979 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.067128] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1117.067128] env[61911]: value = "task-1251684" [ 1117.067128] env[61911]: _type = "Task" [ 1117.067128] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.074260] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.135404] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: da6f7457-e603-415c-9578-024a7563d34f] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1117.319415] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157549} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.319689] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1117.320533] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64043f7d-6c0d-4ad3-8c62-ff8f4f3bcd09 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.342995] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 8ac6fded-77cd-4258-8fcf-c36f27435c20/8ac6fded-77cd-4258-8fcf-c36f27435c20.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.343384] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-344b6b0f-09f9-407d-b296-f1bfbd829038 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.363348] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1117.363348] env[61911]: value = "task-1251685" [ 1117.363348] env[61911]: _type = "Task" [ 1117.363348] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.371374] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251685, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.437414] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5248dd71-3002-3ac5-1c74-7c3680a26113, 'name': SearchDatastore_Task, 'duration_secs': 0.00847} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.438310] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57fad640-0549-4b36-aee4-d15cb2022fff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.443673] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1117.443673] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528c56f5-ef5a-7fd0-da2b-d4407cf6b254" [ 1117.443673] env[61911]: _type = "Task" [ 1117.443673] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.453250] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528c56f5-ef5a-7fd0-da2b-d4407cf6b254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.577120] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251684, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.638969] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 5b7d09e1-ee14-4001-8d73-14e763402670] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1117.873651] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251685, 'name': ReconfigVM_Task, 'duration_secs': 0.293156} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.873988] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 8ac6fded-77cd-4258-8fcf-c36f27435c20/8ac6fded-77cd-4258-8fcf-c36f27435c20.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1117.874685] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-932cc405-beae-48ff-a6a0-ee1f25175bb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.882257] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1117.882257] env[61911]: value = "task-1251686" [ 1117.882257] env[61911]: _type = "Task" [ 1117.882257] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.890582] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251686, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.954186] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]528c56f5-ef5a-7fd0-da2b-d4407cf6b254, 'name': SearchDatastore_Task, 'duration_secs': 0.076362} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.954505] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.954796] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] a3596e80-37b9-4e7b-bf6e-8f995df1deed/a3596e80-37b9-4e7b-bf6e-8f995df1deed.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1117.955181] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cffde3f8-9778-4478-8930-0802a1a935a9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.961939] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1117.961939] env[61911]: value = "task-1251687" [ 1117.961939] env[61911]: _type = "Task" [ 1117.961939] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.970177] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.077633] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251684, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.142680] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: b6b16408-7633-4f23-8f57-1193b1fb0384] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1118.392546] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251686, 'name': Rename_Task, 'duration_secs': 0.19916} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.392920] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1118.393174] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5299989-9fad-4efb-b74b-f9534116cc44 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.400114] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1118.400114] env[61911]: value = "task-1251688" [ 1118.400114] env[61911]: _type = "Task" [ 1118.400114] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.408155] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251688, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.471178] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.578855] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251684, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.645904] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 1058f8d2-0979-4423-802d-52d91d5a2ff7] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1118.910225] env[61911]: DEBUG oslo_vmware.api [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251688, 'name': PowerOnVM_Task, 'duration_secs': 0.468388} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.910469] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1118.910706] env[61911]: INFO nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Took 7.23 seconds to spawn the instance on the hypervisor. [ 1118.910884] env[61911]: DEBUG nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1118.911705] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc57967d-b8ea-4415-8651-d78892085464 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.973896] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.079696] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251684, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.149715] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 9c2e75d8-b307-43bb-84a8-604cd5edbc39] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1119.433698] env[61911]: INFO nova.compute.manager [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Took 14.29 seconds to build instance. [ 1119.474414] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251687, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.580663] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251684, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.260654} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.580964] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7fc0c66b-5933-4158-bc37-1ff7cb26dee0/7fc0c66b-5933-4158-bc37-1ff7cb26dee0.vmdk to [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1119.581870] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7614387e-3d1a-45a8-918a-da295807cf99 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.604604] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.604959] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f00976e-777a-4216-a7d6-bf9cea2cfdf1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.624906] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1119.624906] env[61911]: value = "task-1251689" [ 1119.624906] env[61911]: _type = "Task" [ 1119.624906] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.633651] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251689, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.654639] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: e3ad3416-423c-4687-aa35-2fb2ef67ec64] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1119.736701] env[61911]: DEBUG nova.compute.manager [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Received event network-changed-56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1119.736907] env[61911]: DEBUG nova.compute.manager [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Refreshing instance network info cache due to event network-changed-56626d17-a6ef-4464-b86f-3ca4a0b20e09. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1119.737146] env[61911]: DEBUG oslo_concurrency.lockutils [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] Acquiring lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.737292] env[61911]: DEBUG oslo_concurrency.lockutils [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] Acquired lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.737463] env[61911]: DEBUG nova.network.neutron [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Refreshing network info cache for port 56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1119.936416] env[61911]: DEBUG oslo_concurrency.lockutils [None req-188bbe52-8599-4095-b4aa-c1ad151199f6 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.801s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.973557] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251687, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.746526} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.973831] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] a3596e80-37b9-4e7b-bf6e-8f995df1deed/a3596e80-37b9-4e7b-bf6e-8f995df1deed.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1119.974108] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1119.974362] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9eaf74ff-808d-4165-bf36-2073123b5919 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.979843] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1119.979843] env[61911]: value = "task-1251690" [ 1119.979843] env[61911]: _type = "Task" [ 1119.979843] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.987176] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.135537] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251689, 'name': ReconfigVM_Task, 'duration_secs': 0.341929} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.135822] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b/95853730-1908-42f4-b30f-a4f276d2da8b.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.137181] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encrypted': False, 'encryption_secret_uuid': None, 'boot_index': 0, 'guest_format': None, 'device_name': '/dev/sda', 'encryption_format': None, 'size': 0, 'disk_bus': None, 'encryption_options': None, 'device_type': 'disk', 'image_id': '0dfe2ff1-43fd-4529-93f1-daaccc0711cf'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'boot_index': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269685', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'name': 'volume-f409eb89-364e-4162-a493-56d73a45446b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '95853730-1908-42f4-b30f-a4f276d2da8b', 'attached_at': '', 'detached_at': '', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'serial': 'f409eb89-364e-4162-a493-56d73a45446b'}, 'mount_device': '/dev/sdb', 'disk_bus': None, 'device_type': None, 'attachment_id': '483cd378-c150-4431-a4d6-0ded005e182d', 'volume_type': None}], 'swap': None} {{(pid=61911) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1120.137393] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1120.137589] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269685', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'name': 'volume-f409eb89-364e-4162-a493-56d73a45446b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '95853730-1908-42f4-b30f-a4f276d2da8b', 'attached_at': '', 'detached_at': '', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'serial': 'f409eb89-364e-4162-a493-56d73a45446b'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1120.138525] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6993873-be48-4e9f-af80-e11c4fb8aa8f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.153322] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d16946-df47-420a-9bf7-7080b5d6704a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.172108] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: f0dcfbbf-98ab-4e9f-838f-d9601d2bb045] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1120.181027] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] volume-f409eb89-364e-4162-a493-56d73a45446b/volume-f409eb89-364e-4162-a493-56d73a45446b.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.181452] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8460737-c08d-4241-87cf-0d0887494899 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.199651] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1120.199651] env[61911]: value = "task-1251691" [ 1120.199651] env[61911]: _type = "Task" [ 1120.199651] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.207991] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.489739] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065824} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.490138] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1120.490961] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ade133-f4d0-4e6f-b96d-c037c4e63fdc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.512651] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] a3596e80-37b9-4e7b-bf6e-8f995df1deed/a3596e80-37b9-4e7b-bf6e-8f995df1deed.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.512925] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-400398e6-c29c-4ada-8bae-37b9be000c73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.532419] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1120.532419] env[61911]: value = "task-1251692" [ 1120.532419] env[61911]: _type = "Task" [ 1120.532419] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.540340] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251692, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.681878] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 63d54f69-3d8f-45e9-b317-9d484af954b1] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1120.713021] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251691, 'name': ReconfigVM_Task, 'duration_secs': 0.305799} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.713021] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfigured VM instance instance-00000063 to attach disk [datastore1] volume-f409eb89-364e-4162-a493-56d73a45446b/volume-f409eb89-364e-4162-a493-56d73a45446b.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.716937] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-806440ca-31c5-4170-8556-4e09c6555941 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.730446] env[61911]: DEBUG nova.network.neutron [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updated VIF entry in instance network info cache for port 56626d17-a6ef-4464-b86f-3ca4a0b20e09. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1120.731027] env[61911]: DEBUG nova.network.neutron [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updating instance_info_cache with network_info: [{"id": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "address": "fa:16:3e:45:8f:e1", "network": {"id": "e38e5a4d-3a82-4e5b-b4dd-7291d580c2e5", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1319482734-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e6ca33655884f16b4c0a6fcdb31ec12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56626d17-a6", "ovs_interfaceid": "56626d17-a6ef-4464-b86f-3ca4a0b20e09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.737395] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1120.737395] env[61911]: value = "task-1251693" [ 1120.737395] env[61911]: _type = "Task" [ 1120.737395] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.746109] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251693, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.044477] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251692, 'name': ReconfigVM_Task, 'duration_secs': 0.250223} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.045225] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Reconfigured VM instance instance-0000006a to attach disk [datastore1] a3596e80-37b9-4e7b-bf6e-8f995df1deed/a3596e80-37b9-4e7b-bf6e-8f995df1deed.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.045680] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65c8bd07-39b2-4352-b95b-d3ba08b3a7ae {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.051780] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1121.051780] env[61911]: value = "task-1251694" [ 1121.051780] env[61911]: _type = "Task" [ 1121.051780] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.059703] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251694, 'name': Rename_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.188627] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 9230cd85-960d-427c-b69c-d19ac2bcfe35] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1121.234010] env[61911]: DEBUG oslo_concurrency.lockutils [req-8e84c357-6f35-4569-98cf-343c94733b45 req-3cfbf1bf-f5c6-4131-b6ae-115e29c762a3 service nova] Releasing lock "refresh_cache-8ac6fded-77cd-4258-8fcf-c36f27435c20" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.247046] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251693, 'name': ReconfigVM_Task, 'duration_secs': 0.139384} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.247344] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269685', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'name': 'volume-f409eb89-364e-4162-a493-56d73a45446b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '95853730-1908-42f4-b30f-a4f276d2da8b', 'attached_at': '', 'detached_at': '', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'serial': 'f409eb89-364e-4162-a493-56d73a45446b'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1121.247894] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d69f7e3a-ef4c-4cdc-bdce-de81216bee90 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.253167] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1121.253167] env[61911]: value = "task-1251695" [ 1121.253167] env[61911]: _type = "Task" [ 1121.253167] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.259907] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251695, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.562834] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251694, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.692726] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 4b113b46-1da2-49ad-a0bb-c01b56c70f00] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1121.763558] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251695, 'name': Rename_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.062256] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251694, 'name': Rename_Task} progress is 99%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.196105] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 6e100d92-eb95-4b9b-b236-34ff94051811] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1122.263525] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251695, 'name': Rename_Task} progress is 14%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.563147] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251694, 'name': Rename_Task, 'duration_secs': 1.166517} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.563430] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1122.563698] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ba99dcc-bbd1-4ce3-b440-e0e66ae2a230 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.570237] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1122.570237] env[61911]: value = "task-1251696" [ 1122.570237] env[61911]: _type = "Task" [ 1122.570237] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.577224] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.699293] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 613544bc-92b9-4c8f-8a7b-f02205f347ff] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1122.763669] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251695, 'name': Rename_Task, 'duration_secs': 1.053535} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.763936] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1122.764206] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acf376e1-1ef3-4877-8bda-1073eecbf3d8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.770297] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1122.770297] env[61911]: value = "task-1251697" [ 1122.770297] env[61911]: _type = "Task" [ 1122.770297] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.777436] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251697, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.081978] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251696, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.203096] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 65d310c6-d438-4c3d-bd44-f3b51123fe93] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1123.280539] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251697, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.581503] env[61911]: DEBUG oslo_vmware.api [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251696, 'name': PowerOnVM_Task, 'duration_secs': 0.565282} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.581817] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1123.582047] env[61911]: INFO nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Took 9.77 seconds to spawn the instance on the hypervisor. [ 1123.582242] env[61911]: DEBUG nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1123.583061] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195ef411-bc00-46f6-b3e6-d552b4b43561 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.706433] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 83616524-2530-4398-9019-148e0658720d] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1123.781560] env[61911]: DEBUG oslo_vmware.api [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251697, 'name': PowerOnVM_Task, 'duration_secs': 0.686335} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.781842] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1123.903677] env[61911]: DEBUG nova.compute.manager [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1123.904717] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6145ea-ee8d-48ee-83d0-81dd95eb35c2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.099769] env[61911]: INFO nova.compute.manager [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Took 18.36 seconds to build instance. [ 1124.209582] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: c95707ca-7e32-42c0-914a-f0b178f07997] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1124.422118] env[61911]: DEBUG oslo_concurrency.lockutils [None req-170e8e42-ecd5-451b-9090-25240e5c89b1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 27.253s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.602730] env[61911]: DEBUG oslo_concurrency.lockutils [None req-903cb95f-0882-455b-988f-ebd037da062e tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.875s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.713451] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bafd58bc-8ae7-49a4-a039-31d328699010] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1125.216094] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 0fdf49cd-7a80-4c6d-ba1b-62fbd8e3faaf] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1125.552279] env[61911]: DEBUG nova.compute.manager [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Stashing vm_state: active {{(pid=61911) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 1125.719800] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: e9833b95-4162-42ba-87a4-d4cc790ac8e5] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1126.075428] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.075739] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.223662] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 31f0a4f7-a69d-4ec4-966d-93f402163b88] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1126.580921] env[61911]: INFO nova.compute.claims [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1126.727171] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 43d18895-202c-4048-9435-b3484ffd4c07] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1127.086560] env[61911]: INFO nova.compute.resource_tracker [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating resource usage from migration 7f6c6009-db88-46de-a481-8bc9709e890f [ 1127.146361] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38203cfe-b8c6-43fd-ae4e-eb6f17f6aafe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.153919] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e857601b-53b3-4110-9e33-1cfe487a882b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.182712] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3660ef5a-f560-4aea-9f3e-1c8b36c17c45 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.189571] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb2dcd9-491c-4548-a27f-9cf54c931477 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.203305] env[61911]: DEBUG nova.compute.provider_tree [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.229921] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 28da8a83-a004-4baf-a744-bcc260032afd] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1127.706830] env[61911]: DEBUG nova.scheduler.client.report [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1127.732802] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 951e57ba-da09-426a-b500-e5459d346f64] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1128.212242] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.136s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.212474] env[61911]: INFO nova.compute.manager [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Migrating [ 1128.235176] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 35cf6a24-3cc0-40bd-92da-798a56b4e2c9] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1128.728309] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.728539] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.728731] env[61911]: DEBUG nova.network.neutron [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.738323] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8051e8ec-6d82-416c-858a-f0cdd00f869d] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1129.241399] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8a363b99-517c-45ff-9e33-e9ceff3a08b0] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1129.427355] env[61911]: DEBUG nova.network.neutron [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance_info_cache with network_info: [{"id": "59c51f38-526f-4792-88c9-ba9bb0d19458", "address": "fa:16:3e:c6:e2:6d", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59c51f38-52", "ovs_interfaceid": "59c51f38-526f-4792-88c9-ba9bb0d19458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.744955] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 28159e75-9fe9-44c7-b5c9-534902cecbef] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1129.930629] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.247761] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 267075d7-c3d8-4520-b7d9-67504097f823] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1130.750830] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 66472f43-537d-4eb3-8d49-d40627a8809d] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1131.254637] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 21723e9a-5c6d-4142-baba-25c59a5038ea] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1131.444980] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1baafd61-561f-44b4-b5db-301c5707313f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.463145] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance 'a3596e80-37b9-4e7b-bf6e-8f995df1deed' progress to 0 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1131.969356] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1131.969738] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18d43c52-aa17-47cb-8260-61c5fe8b401e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.976849] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1131.976849] env[61911]: value = "task-1251698" [ 1131.976849] env[61911]: _type = "Task" [ 1131.976849] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.985456] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.115260] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.486990] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251698, 'name': PowerOffVM_Task, 'duration_secs': 0.248015} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.488196] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1132.488196] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance 'a3596e80-37b9-4e7b-bf6e-8f995df1deed' progress to 17 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.994067] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:34Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1132.994483] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1132.994483] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1132.994665] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1132.994827] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1132.995361] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1132.995361] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1132.995538] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1132.995657] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1132.995841] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1132.996037] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1133.001045] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc21e509-79c4-401c-aa90-b4bdd105140c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.016949] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1133.016949] env[61911]: value = "task-1251699" [ 1133.016949] env[61911]: _type = "Task" [ 1133.016949] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.024958] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251699, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.527382] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251699, 'name': ReconfigVM_Task, 'duration_secs': 0.170829} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.527712] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance 'a3596e80-37b9-4e7b-bf6e-8f995df1deed' progress to 33 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1134.034288] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1134.034658] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1134.034694] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1134.034883] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1134.035065] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1134.035227] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1134.035436] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1134.035607] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1134.035785] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1134.035951] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1134.036201] env[61911]: DEBUG nova.virt.hardware [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1134.041454] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1134.041732] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c200e542-7f60-467f-9ba9-938b4ff6535e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.061529] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1134.061529] env[61911]: value = "task-1251700" [ 1134.061529] env[61911]: _type = "Task" [ 1134.061529] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.068714] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251700, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.571342] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251700, 'name': ReconfigVM_Task, 'duration_secs': 0.167574} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.571640] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1134.572396] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f93e2a7-72ff-414e-9e59-e558bf64708f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.595046] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] a3596e80-37b9-4e7b-bf6e-8f995df1deed/a3596e80-37b9-4e7b-bf6e-8f995df1deed.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.595570] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f479ce46-be24-446f-b055-93e9fb3e6043 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.613326] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1134.613326] env[61911]: value = "task-1251701" [ 1134.613326] env[61911]: _type = "Task" [ 1134.613326] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.620670] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251701, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.123013] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251701, 'name': ReconfigVM_Task, 'duration_secs': 0.280133} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.123330] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Reconfigured VM instance instance-0000006a to attach disk [datastore1] a3596e80-37b9-4e7b-bf6e-8f995df1deed/a3596e80-37b9-4e7b-bf6e-8f995df1deed.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.123620] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance 'a3596e80-37b9-4e7b-bf6e-8f995df1deed' progress to 50 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1135.630318] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c945fa-58c9-4684-a8f1-3fc61a5fa774 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.649470] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7b781e-27e5-4889-8116-ab6b0d67f77d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.665798] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance 'a3596e80-37b9-4e7b-bf6e-8f995df1deed' progress to 67 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1136.206398] env[61911]: DEBUG nova.network.neutron [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Port 59c51f38-526f-4792-88c9-ba9bb0d19458 binding to destination host cpu-1 is already ACTIVE {{(pid=61911) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1137.227078] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.227447] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.227538] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.262771] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.263077] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.263164] env[61911]: DEBUG nova.network.neutron [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1139.001419] env[61911]: DEBUG nova.network.neutron [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance_info_cache with network_info: [{"id": "59c51f38-526f-4792-88c9-ba9bb0d19458", "address": "fa:16:3e:c6:e2:6d", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59c51f38-52", "ovs_interfaceid": "59c51f38-526f-4792-88c9-ba9bb0d19458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.503760] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.027571] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ebcf0e5-f1f5-430d-8248-1c656644baf9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.046010] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e991b592-7afd-4fa6-826f-bd8e170001c9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.052478] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance 'a3596e80-37b9-4e7b-bf6e-8f995df1deed' progress to 83 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1140.559069] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1140.559409] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3da0248-8646-493b-a1f1-0e90c430d82c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.567185] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1140.567185] env[61911]: value = "task-1251702" [ 1140.567185] env[61911]: _type = "Task" [ 1140.567185] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.574889] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251702, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.078957] env[61911]: DEBUG oslo_vmware.api [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251702, 'name': PowerOnVM_Task, 'duration_secs': 0.38733} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.079296] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1141.079506] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-6126aff0-574d-4a39-9424-08dcb5fd2eb7 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance 'a3596e80-37b9-4e7b-bf6e-8f995df1deed' progress to 100 {{(pid=61911) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1142.861570] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.862097] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.862097] env[61911]: DEBUG nova.compute.manager [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Going to confirm migration 3 {{(pid=61911) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 1143.424270] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.424472] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquired lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.424660] env[61911]: DEBUG nova.network.neutron [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1143.424851] env[61911]: DEBUG nova.objects.instance [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'info_cache' on Instance uuid a3596e80-37b9-4e7b-bf6e-8f995df1deed {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.634691] env[61911]: DEBUG nova.network.neutron [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance_info_cache with network_info: [{"id": "59c51f38-526f-4792-88c9-ba9bb0d19458", "address": "fa:16:3e:c6:e2:6d", "network": {"id": "1f0350ef-e88a-4ae0-9761-e3c9c6536776", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-866000977-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b285793da304b5e8f9736098ef163c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19440099-773e-4a31-b82e-84a4daa5d8fe", "external-id": "nsx-vlan-transportzone-752", "segmentation_id": 752, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59c51f38-52", "ovs_interfaceid": "59c51f38-526f-4792-88c9-ba9bb0d19458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.137270] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Releasing lock "refresh_cache-a3596e80-37b9-4e7b-bf6e-8f995df1deed" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.137552] env[61911]: DEBUG nova.objects.instance [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lazy-loading 'migration_context' on Instance uuid a3596e80-37b9-4e7b-bf6e-8f995df1deed {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.641068] env[61911]: DEBUG nova.objects.base [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1145.641817] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e58035-5f1a-4622-b20e-ceda36b29f11 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.661207] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e368c0cc-e781-4db6-915c-ac4a33ee086a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.666652] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1145.666652] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a4682c-316f-4562-5ded-10f2c3787923" [ 1145.666652] env[61911]: _type = "Task" [ 1145.666652] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.674984] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a4682c-316f-4562-5ded-10f2c3787923, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.177244] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52a4682c-316f-4562-5ded-10f2c3787923, 'name': SearchDatastore_Task, 'duration_secs': 0.008161} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.177558] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.177798] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.741571] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8445b61-fb2e-4193-bdbf-d1c2b55cd8f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.748950] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a09b887-30a0-49f3-b607-a54573f8cb3f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.777612] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b62240-4f4f-4502-a6a4-e46396120634 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.784438] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fc36d7-2ecb-4fee-bd5c-b1165f7e0991 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.797180] env[61911]: DEBUG nova.compute.provider_tree [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.300523] env[61911]: DEBUG nova.scheduler.client.report [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.311334] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.133s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.868381] env[61911]: INFO nova.scheduler.client.report [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted allocation for migration 7f6c6009-db88-46de-a481-8bc9709e890f [ 1149.374864] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.512s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.887067] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.887413] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.887624] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.887820] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.887994] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.890249] env[61911]: INFO nova.compute.manager [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Terminating instance [ 1151.394071] env[61911]: DEBUG nova.compute.manager [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1151.394337] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1151.395310] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8787aa-0d43-4a53-815f-dc3370c7be6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.403158] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1151.403418] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bdf8294a-20d5-4ad6-9519-9daa4cc91c81 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.409283] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1151.409283] env[61911]: value = "task-1251703" [ 1151.409283] env[61911]: _type = "Task" [ 1151.409283] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.417196] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.918703] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251703, 'name': PowerOffVM_Task, 'duration_secs': 0.221767} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.919165] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1151.919165] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1151.919341] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a68de0bb-9ad0-4485-b683-cd27eac7188c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.978572] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1151.978814] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1151.979013] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleting the datastore file [datastore1] a3596e80-37b9-4e7b-bf6e-8f995df1deed {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1151.979322] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64d81c28-e786-48d5-98e7-91f4081d8184 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.986023] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for the task: (returnval){ [ 1151.986023] env[61911]: value = "task-1251705" [ 1151.986023] env[61911]: _type = "Task" [ 1151.986023] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.993337] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.495879] env[61911]: DEBUG oslo_vmware.api [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Task: {'id': task-1251705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14197} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.496173] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1152.496371] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1152.496560] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1152.496780] env[61911]: INFO nova.compute.manager [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1152.497064] env[61911]: DEBUG oslo.service.loopingcall [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1152.497272] env[61911]: DEBUG nova.compute.manager [-] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1152.497374] env[61911]: DEBUG nova.network.neutron [-] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1152.738238] env[61911]: DEBUG nova.compute.manager [req-1fb78aaa-8e1d-44ad-a07b-85eebc723833 req-0bde322e-fd23-4b6c-ac9e-51654a276a2a service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Received event network-vif-deleted-59c51f38-526f-4792-88c9-ba9bb0d19458 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1152.738299] env[61911]: INFO nova.compute.manager [req-1fb78aaa-8e1d-44ad-a07b-85eebc723833 req-0bde322e-fd23-4b6c-ac9e-51654a276a2a service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Neutron deleted interface 59c51f38-526f-4792-88c9-ba9bb0d19458; detaching it from the instance and deleting it from the info cache [ 1152.738477] env[61911]: DEBUG nova.network.neutron [req-1fb78aaa-8e1d-44ad-a07b-85eebc723833 req-0bde322e-fd23-4b6c-ac9e-51654a276a2a service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.218732] env[61911]: DEBUG nova.network.neutron [-] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.240884] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ff9c3c9-15be-4f50-bcd8-8e26339b69d0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.250352] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e1dcaa-d863-481d-a688-3d9bd3bb7def {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.274024] env[61911]: DEBUG nova.compute.manager [req-1fb78aaa-8e1d-44ad-a07b-85eebc723833 req-0bde322e-fd23-4b6c-ac9e-51654a276a2a service nova] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Detach interface failed, port_id=59c51f38-526f-4792-88c9-ba9bb0d19458, reason: Instance a3596e80-37b9-4e7b-bf6e-8f995df1deed could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1153.721846] env[61911]: INFO nova.compute.manager [-] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Took 1.22 seconds to deallocate network for instance. [ 1153.768826] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1154.229359] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.229694] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.229837] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.250973] env[61911]: INFO nova.scheduler.client.report [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Deleted allocations for instance a3596e80-37b9-4e7b-bf6e-8f995df1deed [ 1154.271785] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Getting list of instances from cluster (obj){ [ 1154.271785] env[61911]: value = "domain-c8" [ 1154.271785] env[61911]: _type = "ClusterComputeResource" [ 1154.271785] env[61911]: } {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1154.272820] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7acd34-ad05-4e3f-abf3-847720e7cb67 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.284936] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Got total of 2 instances {{(pid=61911) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1154.285100] env[61911]: WARNING nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] While synchronizing instance power states, found 3 instances in the database and 2 instances on the hypervisor. [ 1154.285253] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1154.285478] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid 8ac6fded-77cd-4258-8fcf-c36f27435c20 {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1154.285592] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Triggering sync for uuid a3596e80-37b9-4e7b-bf6e-8f995df1deed {{(pid=61911) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1154.285915] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.286142] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.286453] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.286581] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.286867] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.287631] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6978a98-b9c9-4a8b-b911-b2fd0f8620d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.290419] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00518daa-114a-4471-8269-ab2f2e27f7a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.737565] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1154.759458] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0d8595ad-862f-4937-8b87-dc76d11b3294 tempest-DeleteServersTestJSON-1532949826 tempest-DeleteServersTestJSON-1532949826-project-member] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.872s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.760339] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.473s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.761086] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-965b0258-c789-474b-b55d-1adbc097c92d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.770056] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b521b1-563c-44ba-bac0-4fde6752a4dc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.801476] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.515s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.802138] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.515s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.240955] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.241343] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.241386] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.241531] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1155.242438] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5cc79f4-8c20-4b1e-a97d-3d9d882c75ff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.250530] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c62c24-cfb1-4647-87bd-029e7b371887 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.264626] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95064ace-2592-4ed2-95eb-4c7f19036854 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.270872] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb6b4a2-1d74-4362-ad07-101c7534a534 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.298727] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "a3596e80-37b9-4e7b-bf6e-8f995df1deed" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.538s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.299018] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181016MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1155.299160] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.299377] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.057813] env[61911]: DEBUG oslo_concurrency.lockutils [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.058014] env[61911]: DEBUG oslo_concurrency.lockutils [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.323418] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 95853730-1908-42f4-b30f-a4f276d2da8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1156.323687] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 8ac6fded-77cd-4258-8fcf-c36f27435c20 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1156.323785] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1156.323927] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1156.357063] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ff4e22-b21d-4566-94bb-eb113607e7d9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.365783] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82eb0831-2e75-40e6-9110-7714a21699b6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.393800] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6c8dfa-6a87-4168-a8b6-1c100f90e17f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.400429] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efed47b2-89fe-44e0-b3a4-8926ab5d0da5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.412644] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.561763] env[61911]: DEBUG nova.compute.utils [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.915933] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1157.064641] env[61911]: DEBUG oslo_concurrency.lockutils [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.420887] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1157.421289] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.122s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.123946] env[61911]: DEBUG oslo_concurrency.lockutils [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.124255] env[61911]: DEBUG oslo_concurrency.lockutils [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.124517] env[61911]: INFO nova.compute.manager [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Attaching volume a1528900-1dc5-49c7-9b87-5301269a1bea to /dev/sdb [ 1158.153888] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad47f91-87f1-49c7-b09e-5a7724ad7263 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.162274] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f11527c-ca93-4f96-a0bf-65601302fb2a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.174857] env[61911]: DEBUG nova.virt.block_device [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updating existing volume attachment record: c09b621f-9a94-4752-9003-2ea1f33d19f8 {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1159.415079] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.415353] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.415499] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.415653] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.416111] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.737270] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.737451] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1160.728792] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "2077036f-3451-416c-8783-aa3ea5fbcbcb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.729071] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "2077036f-3451-416c-8783-aa3ea5fbcbcb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.737213] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.737376] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1161.140758] env[61911]: DEBUG oslo_concurrency.lockutils [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.141065] env[61911]: DEBUG oslo_concurrency.lockutils [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.231868] env[61911]: DEBUG nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1161.240263] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Didn't find any instances for network info cache update. {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 1161.644407] env[61911]: INFO nova.compute.manager [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Detaching volume f409eb89-364e-4162-a493-56d73a45446b [ 1161.674460] env[61911]: INFO nova.virt.block_device [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Attempting to driver detach volume f409eb89-364e-4162-a493-56d73a45446b from mountpoint /dev/sdb [ 1161.674511] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1161.675111] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269685', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'name': 'volume-f409eb89-364e-4162-a493-56d73a45446b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '95853730-1908-42f4-b30f-a4f276d2da8b', 'attached_at': '', 'detached_at': '', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'serial': 'f409eb89-364e-4162-a493-56d73a45446b'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1161.675701] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb77a88-c367-4a70-96ec-8dea320f59c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.696795] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31682f3b-e074-41d9-b76e-5861abd7ff73 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.703551] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08be1ab-f184-458a-b55a-e5a66c869f94 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.724116] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55620e9d-aa6d-4c91-b506-4b12e07906af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.741814] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] The volume has not been displaced from its original location: [datastore1] volume-f409eb89-364e-4162-a493-56d73a45446b/volume-f409eb89-364e-4162-a493-56d73a45446b.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1161.746956] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1161.747393] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cd79d2f-8109-486f-866d-d9ea8b5bf67e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.761287] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.761535] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.763030] env[61911]: INFO nova.compute.claims [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1161.771407] env[61911]: DEBUG oslo_vmware.api [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1161.771407] env[61911]: value = "task-1251709" [ 1161.771407] env[61911]: _type = "Task" [ 1161.771407] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.778737] env[61911]: DEBUG oslo_vmware.api [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251709, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.281792] env[61911]: DEBUG oslo_vmware.api [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251709, 'name': ReconfigVM_Task, 'duration_secs': 0.218554} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.281792] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1162.286170] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dae431df-dd77-4365-9b25-a08be64650fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.300676] env[61911]: DEBUG oslo_vmware.api [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1162.300676] env[61911]: value = "task-1251710" [ 1162.300676] env[61911]: _type = "Task" [ 1162.300676] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.310828] env[61911]: DEBUG oslo_vmware.api [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.717945] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1162.718212] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269690', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'name': 'volume-a1528900-1dc5-49c7-9b87-5301269a1bea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8ac6fded-77cd-4258-8fcf-c36f27435c20', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'serial': 'a1528900-1dc5-49c7-9b87-5301269a1bea'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1162.719217] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5f2ea7-5c17-43e7-9ba3-dd9acd2a5b32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.735560] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c3aed9-88d2-417b-a4a6-d97251de50c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.738077] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.759245] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] volume-a1528900-1dc5-49c7-9b87-5301269a1bea/volume-a1528900-1dc5-49c7-9b87-5301269a1bea.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.759542] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6e22c86-01bc-456c-ac66-7ba72e9600a2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.778734] env[61911]: DEBUG oslo_vmware.api [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1162.778734] env[61911]: value = "task-1251711" [ 1162.778734] env[61911]: _type = "Task" [ 1162.778734] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.786759] env[61911]: DEBUG oslo_vmware.api [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251711, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.812623] env[61911]: DEBUG oslo_vmware.api [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251710, 'name': ReconfigVM_Task, 'duration_secs': 0.136923} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.815414] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269685', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'name': 'volume-f409eb89-364e-4162-a493-56d73a45446b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '95853730-1908-42f4-b30f-a4f276d2da8b', 'attached_at': '', 'detached_at': '', 'volume_id': 'f409eb89-364e-4162-a493-56d73a45446b', 'serial': 'f409eb89-364e-4162-a493-56d73a45446b'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1162.827529] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7304c22f-9116-4b87-af5e-3258add80956 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.834709] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8461a58-8a78-4dca-aede-551be1351f34 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.863569] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b014bfe9-5ca6-4f26-b39c-1a7dd87949ec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.870022] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28593044-a215-4776-8879-1e9192462d5f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.882484] env[61911]: DEBUG nova.compute.provider_tree [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.288646] env[61911]: DEBUG oslo_vmware.api [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251711, 'name': ReconfigVM_Task, 'duration_secs': 0.328173} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.288969] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Reconfigured VM instance instance-00000069 to attach disk [datastore1] volume-a1528900-1dc5-49c7-9b87-5301269a1bea/volume-a1528900-1dc5-49c7-9b87-5301269a1bea.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.293798] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f134fc00-e067-4e5b-b08c-0618dd633aa8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.308147] env[61911]: DEBUG oslo_vmware.api [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1163.308147] env[61911]: value = "task-1251712" [ 1163.308147] env[61911]: _type = "Task" [ 1163.308147] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.315777] env[61911]: DEBUG oslo_vmware.api [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251712, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.354155] env[61911]: DEBUG nova.objects.instance [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'flavor' on Instance uuid 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.385579] env[61911]: DEBUG nova.scheduler.client.report [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1163.818405] env[61911]: DEBUG oslo_vmware.api [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251712, 'name': ReconfigVM_Task, 'duration_secs': 0.147119} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.818759] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269690', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'name': 'volume-a1528900-1dc5-49c7-9b87-5301269a1bea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8ac6fded-77cd-4258-8fcf-c36f27435c20', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'serial': 'a1528900-1dc5-49c7-9b87-5301269a1bea'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1163.889897] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.128s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.890459] env[61911]: DEBUG nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1164.360401] env[61911]: DEBUG oslo_concurrency.lockutils [None req-dcb3f1c3-f93d-4c1a-849a-8525bcb1bcb1 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.219s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.395174] env[61911]: DEBUG nova.compute.utils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1164.396556] env[61911]: DEBUG nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Not allocating networking since 'none' was specified. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1164.854901] env[61911]: DEBUG nova.objects.instance [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'flavor' on Instance uuid 8ac6fded-77cd-4258-8fcf-c36f27435c20 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.897740] env[61911]: DEBUG nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1165.359371] env[61911]: DEBUG oslo_concurrency.lockutils [None req-05dac972-6db4-40a9-a349-80aac0e97843 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.381970] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.382394] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.382475] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.382639] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.382812] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.384772] env[61911]: INFO nova.compute.manager [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Terminating instance [ 1165.552802] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.552956] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.888611] env[61911]: DEBUG nova.compute.manager [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1165.888984] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1165.889838] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215e06c3-7eab-4498-ad8c-1d0896671838 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.897186] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.897415] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-444399c9-848f-4e53-9f3e-5fa8a4f0572e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.903398] env[61911]: DEBUG oslo_vmware.api [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1165.903398] env[61911]: value = "task-1251713" [ 1165.903398] env[61911]: _type = "Task" [ 1165.903398] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.907217] env[61911]: DEBUG nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1165.913818] env[61911]: DEBUG oslo_vmware.api [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.935783] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1165.936047] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1165.936216] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1165.936416] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1165.936659] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1165.936898] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1165.937189] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1165.937393] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1165.937610] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1165.937823] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1165.938057] env[61911]: DEBUG nova.virt.hardware [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1165.938957] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af5a215-a4a4-43f3-94ef-cd0463cdc775 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.946625] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da9a841-d31b-424f-b22c-defa31cc9fb9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.960960] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1165.966565] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Creating folder: Project (efbaf75390a447ba8e965c5e0013d749). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1165.966877] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ee0a99c-4842-462d-a87f-a359d4fa8aa0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.976754] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Created folder: Project (efbaf75390a447ba8e965c5e0013d749) in parent group-v269521. [ 1165.976960] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Creating folder: Instances. Parent ref: group-v269691. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1165.977172] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f77517e-cbbe-4176-b94c-13a44fd2c48b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.985570] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Created folder: Instances in parent group-v269691. [ 1165.985798] env[61911]: DEBUG oslo.service.loopingcall [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.985979] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1165.986188] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d71ec3ba-5fff-40f4-a6b8-187f6052e87d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.001927] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1166.001927] env[61911]: value = "task-1251716" [ 1166.001927] env[61911]: _type = "Task" [ 1166.001927] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.008994] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251716, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.056177] env[61911]: INFO nova.compute.manager [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Detaching volume a1528900-1dc5-49c7-9b87-5301269a1bea [ 1166.094021] env[61911]: INFO nova.virt.block_device [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Attempting to driver detach volume a1528900-1dc5-49c7-9b87-5301269a1bea from mountpoint /dev/sdb [ 1166.094021] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1166.094181] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269690', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'name': 'volume-a1528900-1dc5-49c7-9b87-5301269a1bea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8ac6fded-77cd-4258-8fcf-c36f27435c20', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'serial': 'a1528900-1dc5-49c7-9b87-5301269a1bea'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1166.095234] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c6cb28-58c6-4402-b122-32ceebb2b7b4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.118294] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edb879d-526a-433f-a4d6-8f1e247baaa6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.125579] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c523a0-e0ed-4a6b-b40d-4d499e036057 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.145753] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe3ab34-9c30-4a4d-b660-8444cd1ee5db {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.161897] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] The volume has not been displaced from its original location: [datastore1] volume-a1528900-1dc5-49c7-9b87-5301269a1bea/volume-a1528900-1dc5-49c7-9b87-5301269a1bea.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1166.167594] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1166.167869] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae92c60c-3638-4a71-8a5b-16cdc896404c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.185698] env[61911]: DEBUG oslo_vmware.api [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1166.185698] env[61911]: value = "task-1251717" [ 1166.185698] env[61911]: _type = "Task" [ 1166.185698] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.193588] env[61911]: DEBUG oslo_vmware.api [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251717, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.412465] env[61911]: DEBUG oslo_vmware.api [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251713, 'name': PowerOffVM_Task, 'duration_secs': 0.196594} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.412719] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1166.412898] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1166.413159] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-637de1b9-9f53-4a10-80d5-15049de539ec {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.471853] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.472122] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.472367] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleting the datastore file [datastore1] 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.472651] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a43f984d-2bfc-48e5-85de-bbd9805e96ef {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.478111] env[61911]: DEBUG oslo_vmware.api [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1166.478111] env[61911]: value = "task-1251719" [ 1166.478111] env[61911]: _type = "Task" [ 1166.478111] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.484967] env[61911]: DEBUG oslo_vmware.api [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251719, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.511833] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251716, 'name': CreateVM_Task, 'duration_secs': 0.242929} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.512034] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1166.512534] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.512703] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.513053] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1166.513313] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-772fc124-a487-464c-817c-3850120095e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.517709] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1166.517709] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c880fd-4144-7793-a6ac-bb0a1c98e6be" [ 1166.517709] env[61911]: _type = "Task" [ 1166.517709] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.524676] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c880fd-4144-7793-a6ac-bb0a1c98e6be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.695661] env[61911]: DEBUG oslo_vmware.api [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251717, 'name': ReconfigVM_Task, 'duration_secs': 0.206033} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.695906] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1166.700413] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a97fe61-f5bc-4481-9d69-fc3ce62dfa75 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.714919] env[61911]: DEBUG oslo_vmware.api [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1166.714919] env[61911]: value = "task-1251720" [ 1166.714919] env[61911]: _type = "Task" [ 1166.714919] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.722104] env[61911]: DEBUG oslo_vmware.api [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251720, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.988853] env[61911]: DEBUG oslo_vmware.api [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251719, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122686} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.989303] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1166.989370] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1166.989497] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1166.989684] env[61911]: INFO nova.compute.manager [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1166.990059] env[61911]: DEBUG oslo.service.loopingcall [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1166.990358] env[61911]: DEBUG nova.compute.manager [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1166.990536] env[61911]: DEBUG nova.network.neutron [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1167.030079] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c880fd-4144-7793-a6ac-bb0a1c98e6be, 'name': SearchDatastore_Task, 'duration_secs': 0.017389} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.031744] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.031996] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1167.032393] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.032579] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.032770] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1167.033043] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1c890c0-68d5-444a-9b55-3d87a31eadac {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.041052] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1167.041052] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1167.041563] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65064b88-93d1-46bf-bd7f-376b21a36ab1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.046272] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1167.046272] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5242eea6-7b58-1a91-e186-4fa84724c03b" [ 1167.046272] env[61911]: _type = "Task" [ 1167.046272] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.053700] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5242eea6-7b58-1a91-e186-4fa84724c03b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.224581] env[61911]: DEBUG oslo_vmware.api [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251720, 'name': ReconfigVM_Task, 'duration_secs': 0.124548} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.224936] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269690', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'name': 'volume-a1528900-1dc5-49c7-9b87-5301269a1bea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8ac6fded-77cd-4258-8fcf-c36f27435c20', 'attached_at': '', 'detached_at': '', 'volume_id': 'a1528900-1dc5-49c7-9b87-5301269a1bea', 'serial': 'a1528900-1dc5-49c7-9b87-5301269a1bea'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1167.512375] env[61911]: DEBUG nova.compute.manager [req-b7924ed2-e543-46e5-be93-f6375fa7599c req-1db20bb0-c64a-4cd1-89b4-d64154340975 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Received event network-vif-deleted-19140385-696d-4e4c-a2dd-d53c936d4416 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1167.513421] env[61911]: INFO nova.compute.manager [req-b7924ed2-e543-46e5-be93-f6375fa7599c req-1db20bb0-c64a-4cd1-89b4-d64154340975 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Neutron deleted interface 19140385-696d-4e4c-a2dd-d53c936d4416; detaching it from the instance and deleting it from the info cache [ 1167.513828] env[61911]: DEBUG nova.network.neutron [req-b7924ed2-e543-46e5-be93-f6375fa7599c req-1db20bb0-c64a-4cd1-89b4-d64154340975 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.556437] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5242eea6-7b58-1a91-e186-4fa84724c03b, 'name': SearchDatastore_Task, 'duration_secs': 0.007731} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.557481] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3bd64c0-bebd-4170-9c69-29b1ec869ae9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.562371] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1167.562371] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d0baf1-bd88-37d7-84f6-4ac77b09d686" [ 1167.562371] env[61911]: _type = "Task" [ 1167.562371] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.569713] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d0baf1-bd88-37d7-84f6-4ac77b09d686, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.766912] env[61911]: DEBUG nova.objects.instance [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'flavor' on Instance uuid 8ac6fded-77cd-4258-8fcf-c36f27435c20 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.960027] env[61911]: DEBUG nova.network.neutron [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.018660] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bda551b6-62e1-4b69-b751-78c93d1ff0aa {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.028558] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992cd2b0-98e0-4e30-aab8-db109512c32e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.054207] env[61911]: DEBUG nova.compute.manager [req-b7924ed2-e543-46e5-be93-f6375fa7599c req-1db20bb0-c64a-4cd1-89b4-d64154340975 service nova] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Detach interface failed, port_id=19140385-696d-4e4c-a2dd-d53c936d4416, reason: Instance 95853730-1908-42f4-b30f-a4f276d2da8b could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1168.071843] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d0baf1-bd88-37d7-84f6-4ac77b09d686, 'name': SearchDatastore_Task, 'duration_secs': 0.008555} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.072109] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.072398] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1168.072644] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a00e1187-6c38-4fd3-abba-9023d23a0741 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.078686] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1168.078686] env[61911]: value = "task-1251721" [ 1168.078686] env[61911]: _type = "Task" [ 1168.078686] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.085494] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251721, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.462788] env[61911]: INFO nova.compute.manager [-] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Took 1.47 seconds to deallocate network for instance. [ 1168.587733] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251721, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424259} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.588015] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1168.588247] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1168.588491] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be27ad92-ce97-420f-bd4b-426901ce328b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.595097] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1168.595097] env[61911]: value = "task-1251722" [ 1168.595097] env[61911]: _type = "Task" [ 1168.595097] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.601914] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251722, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.774873] env[61911]: DEBUG oslo_concurrency.lockutils [None req-ba68524d-2104-4187-af67-9e7d0bb563ba tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.222s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.969387] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.969611] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.969839] env[61911]: DEBUG nova.objects.instance [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'resources' on Instance uuid 95853730-1908-42f4-b30f-a4f276d2da8b {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.105091] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251722, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067969} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.105466] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1169.106151] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d38782-543f-4249-bd2c-6f6b894755f5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.125248] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1169.125473] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a962505-1e13-4695-ae9c-d5f9bd529992 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.143605] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1169.143605] env[61911]: value = "task-1251723" [ 1169.143605] env[61911]: _type = "Task" [ 1169.143605] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.150568] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251723, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.524829] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01ac1d6-f34d-427a-8203-ff62499a07fb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.532094] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676fc6d5-b43b-48ba-a05b-8c0422aa67ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.561284] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff31462-b4ed-4ea6-a4d3-0ec252b5df47 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.567927] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8857e6b2-2780-4a9f-862e-fa2c1f81896e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.580300] env[61911]: DEBUG nova.compute.provider_tree [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.652066] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251723, 'name': ReconfigVM_Task, 'duration_secs': 0.293623} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.652352] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1169.653178] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96b2a624-1a57-441c-b9a8-0ad32b2d9237 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.659174] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1169.659174] env[61911]: value = "task-1251724" [ 1169.659174] env[61911]: _type = "Task" [ 1169.659174] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.666421] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251724, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.791758] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.791985] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.792248] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "8ac6fded-77cd-4258-8fcf-c36f27435c20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.792453] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.792632] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.794781] env[61911]: INFO nova.compute.manager [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Terminating instance [ 1170.083728] env[61911]: DEBUG nova.scheduler.client.report [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1170.170134] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251724, 'name': Rename_Task, 'duration_secs': 0.127391} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.170472] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1170.170644] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26373d51-b0ba-46d5-a7c0-299f67bdf18f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.176299] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1170.176299] env[61911]: value = "task-1251725" [ 1170.176299] env[61911]: _type = "Task" [ 1170.176299] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.188486] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.298083] env[61911]: DEBUG nova.compute.manager [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1170.298341] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1170.299312] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e556d2-afae-4d0c-a8f9-49a5462bf43e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.306655] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1170.306896] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78ca4256-f115-4c8e-a0bd-19162eb18741 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.312987] env[61911]: DEBUG oslo_vmware.api [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1170.312987] env[61911]: value = "task-1251726" [ 1170.312987] env[61911]: _type = "Task" [ 1170.312987] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.320170] env[61911]: DEBUG oslo_vmware.api [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251726, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.588766] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.606301] env[61911]: INFO nova.scheduler.client.report [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted allocations for instance 95853730-1908-42f4-b30f-a4f276d2da8b [ 1170.686269] env[61911]: DEBUG oslo_vmware.api [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251725, 'name': PowerOnVM_Task, 'duration_secs': 0.406326} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.686489] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1170.686693] env[61911]: INFO nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Took 4.78 seconds to spawn the instance on the hypervisor. [ 1170.687045] env[61911]: DEBUG nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1170.687614] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7faa4f9a-4290-454a-a383-064af1ba44b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.822942] env[61911]: DEBUG oslo_vmware.api [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251726, 'name': PowerOffVM_Task, 'duration_secs': 0.175679} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.823254] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1170.823440] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1170.823684] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9e97c3d-7336-4ffb-b507-7fad7fcbf772 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.885483] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1170.885710] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.885900] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleting the datastore file [datastore1] 8ac6fded-77cd-4258-8fcf-c36f27435c20 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.886170] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85817f5d-814c-4062-ae17-540694c496bb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.892613] env[61911]: DEBUG oslo_vmware.api [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for the task: (returnval){ [ 1170.892613] env[61911]: value = "task-1251728" [ 1170.892613] env[61911]: _type = "Task" [ 1170.892613] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.899999] env[61911]: DEBUG oslo_vmware.api [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.114313] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8ab01de1-ee6e-4279-8cdc-47fabb0ee6db tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "95853730-1908-42f4-b30f-a4f276d2da8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.732s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.202303] env[61911]: INFO nova.compute.manager [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Took 9.46 seconds to build instance. [ 1171.402995] env[61911]: DEBUG oslo_vmware.api [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Task: {'id': task-1251728, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136735} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.403234] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.403434] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1171.403616] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1171.403794] env[61911]: INFO nova.compute.manager [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1171.404047] env[61911]: DEBUG oslo.service.loopingcall [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.404246] env[61911]: DEBUG nova.compute.manager [-] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1171.404341] env[61911]: DEBUG nova.network.neutron [-] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1171.704275] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8f18fc7c-ddfe-4f06-86d8-f7fbf0debede tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "2077036f-3451-416c-8783-aa3ea5fbcbcb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 10.975s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.960735] env[61911]: DEBUG nova.compute.manager [req-b58a5fdc-04ec-4ac7-8712-bf63b964a95d req-69e0a827-b4d1-4ce8-aab1-c5151bd6d7a9 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Received event network-vif-deleted-56626d17-a6ef-4464-b86f-3ca4a0b20e09 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1171.964018] env[61911]: INFO nova.compute.manager [req-b58a5fdc-04ec-4ac7-8712-bf63b964a95d req-69e0a827-b4d1-4ce8-aab1-c5151bd6d7a9 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Neutron deleted interface 56626d17-a6ef-4464-b86f-3ca4a0b20e09; detaching it from the instance and deleting it from the info cache [ 1171.964018] env[61911]: DEBUG nova.network.neutron [req-b58a5fdc-04ec-4ac7-8712-bf63b964a95d req-69e0a827-b4d1-4ce8-aab1-c5151bd6d7a9 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.438288] env[61911]: DEBUG nova.network.neutron [-] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.463131] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d430e906-03e1-448b-92c7-cc007b45280d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.473035] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe7dcf4-ba8a-42d4-8e92-b3d54d0a1d9a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.497417] env[61911]: INFO nova.compute.manager [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Rebuilding instance [ 1172.499291] env[61911]: DEBUG nova.compute.manager [req-b58a5fdc-04ec-4ac7-8712-bf63b964a95d req-69e0a827-b4d1-4ce8-aab1-c5151bd6d7a9 service nova] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Detach interface failed, port_id=56626d17-a6ef-4464-b86f-3ca4a0b20e09, reason: Instance 8ac6fded-77cd-4258-8fcf-c36f27435c20 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1172.542960] env[61911]: DEBUG nova.compute.manager [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1172.543883] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b8390c-682c-47e8-93ad-a3a5bfc114cb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.943037] env[61911]: INFO nova.compute.manager [-] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Took 1.54 seconds to deallocate network for instance. [ 1173.448388] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.448831] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.448892] env[61911]: DEBUG nova.objects.instance [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lazy-loading 'resources' on Instance uuid 8ac6fded-77cd-4258-8fcf-c36f27435c20 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.559073] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1173.559450] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-359bd83a-0f3a-4252-a63c-83af602bc4a6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.566757] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1173.566757] env[61911]: value = "task-1251729" [ 1173.566757] env[61911]: _type = "Task" [ 1173.566757] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.575294] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.636562] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.636852] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.999984] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04686827-72b0-4e86-9fbd-e7ee1ece1701 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.007616] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b1edbb-a123-4f3f-9628-4e2eb424d43b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.037707] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7e4d85-3d7f-4c6b-845b-d7fe99559801 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.044397] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524be259-45da-4600-87b2-19e63eff5597 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.056878] env[61911]: DEBUG nova.compute.provider_tree [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.075445] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251729, 'name': PowerOffVM_Task, 'duration_secs': 0.184504} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.075685] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.075921] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1174.076618] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd11fec-b379-40f4-a1dc-9f03749ce0e2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.082546] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1174.082751] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6da9e25c-8823-4e2d-88ab-1c4b19359719 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.103996] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1174.104229] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1174.104427] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Deleting the datastore file [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.104663] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfcf26c0-6cd9-4e8d-a064-4adc01d49472 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.111342] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1174.111342] env[61911]: value = "task-1251731" [ 1174.111342] env[61911]: _type = "Task" [ 1174.111342] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.118227] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.139172] env[61911]: DEBUG nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1174.560761] env[61911]: DEBUG nova.scheduler.client.report [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1174.620867] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10876} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.621152] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1174.621371] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1174.621567] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1174.656047] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.065884] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.068176] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.412s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.070425] env[61911]: INFO nova.compute.claims [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1175.084362] env[61911]: INFO nova.scheduler.client.report [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Deleted allocations for instance 8ac6fded-77cd-4258-8fcf-c36f27435c20 [ 1175.590629] env[61911]: DEBUG oslo_concurrency.lockutils [None req-81a85631-9af6-4521-8697-3e0ecee71f60 tempest-AttachVolumeNegativeTest-633915258 tempest-AttachVolumeNegativeTest-633915258-project-member] Lock "8ac6fded-77cd-4258-8fcf-c36f27435c20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.798s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.653705] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1175.653955] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1175.654134] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1175.654329] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1175.654484] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1175.654680] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1175.654905] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1175.655089] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1175.655267] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1175.655444] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1175.655617] env[61911]: DEBUG nova.virt.hardware [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1175.656503] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fdd5a5-7df1-42f8-9c9f-e837e738ec7a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.664431] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ae7ec0-75f4-42e7-9a52-6d029db21080 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.677878] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Instance VIF info [] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1175.683518] env[61911]: DEBUG oslo.service.loopingcall [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1175.683744] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1175.683940] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a82ae8e-eac7-4ea6-be76-c1fb36ea1945 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.700870] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1175.700870] env[61911]: value = "task-1251732" [ 1175.700870] env[61911]: _type = "Task" [ 1175.700870] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.707960] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251732, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.115388] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b31396-50a2-4fd5-a7f3-5b597c7b178d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.123239] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee850cc-451c-4a91-9eea-fad3be22c154 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.152150] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64776b85-b614-41c0-ba52-2f6293a48b96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.159429] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35819c9a-959a-43aa-a005-1bd92615ff7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.173944] env[61911]: DEBUG nova.compute.provider_tree [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.215649] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251732, 'name': CreateVM_Task, 'duration_secs': 0.239751} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.215835] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1176.216295] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.216468] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.216836] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1176.217951] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06ada7cb-ffe4-463e-880d-3028848dbe67 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.223293] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1176.223293] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c72682-6bd6-b664-9317-ea63e5cf94ca" [ 1176.223293] env[61911]: _type = "Task" [ 1176.223293] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.229492] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c72682-6bd6-b664-9317-ea63e5cf94ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.676716] env[61911]: DEBUG nova.scheduler.client.report [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1176.731963] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c72682-6bd6-b664-9317-ea63e5cf94ca, 'name': SearchDatastore_Task, 'duration_secs': 0.009498} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.732317] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.732556] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1176.732795] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.732951] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.733148] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.733399] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38a9bac7-47f0-4cc5-ac78-af8b5e7643d0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.740858] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.741043] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1176.741721] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a9a7e2b-9aa9-4fcf-aa1a-7f8e48289c92 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.746483] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1176.746483] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ebbe4e-0c65-fdfe-ef2a-814b044d9678" [ 1176.746483] env[61911]: _type = "Task" [ 1176.746483] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.753253] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ebbe4e-0c65-fdfe-ef2a-814b044d9678, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.181901] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.182395] env[61911]: DEBUG nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1177.256399] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52ebbe4e-0c65-fdfe-ef2a-814b044d9678, 'name': SearchDatastore_Task, 'duration_secs': 0.007629} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.257178] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d23aff26-aab7-47a8-ac4a-7cafe66a2380 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.261901] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1177.261901] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c46f49-c7b5-e630-5899-5856ee074890" [ 1177.261901] env[61911]: _type = "Task" [ 1177.261901] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.268803] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c46f49-c7b5-e630-5899-5856ee074890, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.687333] env[61911]: DEBUG nova.compute.utils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1177.688757] env[61911]: DEBUG nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1177.688950] env[61911]: DEBUG nova.network.neutron [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1177.729433] env[61911]: DEBUG nova.policy [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc117e32967d4861a77259c18e0c9c33', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce8675694c2841a58e87bb7250f3435a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1177.771877] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52c46f49-c7b5-e630-5899-5856ee074890, 'name': SearchDatastore_Task, 'duration_secs': 0.011142} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.773538] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.773799] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1177.774071] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60ad22e3-713a-4899-9257-2b466ffeb1a2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.781686] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1177.781686] env[61911]: value = "task-1251734" [ 1177.781686] env[61911]: _type = "Task" [ 1177.781686] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.789570] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251734, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.007299] env[61911]: DEBUG nova.network.neutron [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Successfully created port: cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1178.192575] env[61911]: DEBUG nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1178.291583] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251734, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442896} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.291855] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1178.292116] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1178.292383] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cc90b06-2099-4c77-a3da-3aa26115c7b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.298524] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1178.298524] env[61911]: value = "task-1251735" [ 1178.298524] env[61911]: _type = "Task" [ 1178.298524] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.305692] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251735, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.807804] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251735, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063873} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.808106] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1178.808850] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c591fb0e-6f50-426d-9fc1-27127302cd23 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.827970] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1178.828216] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db49c309-26e7-4d6c-990b-7496ccae31be {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.846240] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1178.846240] env[61911]: value = "task-1251736" [ 1178.846240] env[61911]: _type = "Task" [ 1178.846240] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.853449] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251736, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.203025] env[61911]: DEBUG nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1179.228113] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1179.228378] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1179.228542] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1179.228734] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1179.228885] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1179.229050] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1179.229304] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1179.229478] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1179.229650] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1179.229818] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1179.229998] env[61911]: DEBUG nova.virt.hardware [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1179.230873] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1323c7e9-3dcf-49e5-88a5-57feeb499b7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.238917] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323231e5-283a-449e-9892-d3e5b234799f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.355785] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251736, 'name': ReconfigVM_Task, 'duration_secs': 0.257017} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.356085] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb/2077036f-3451-416c-8783-aa3ea5fbcbcb.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1179.356704] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de335502-4745-4f3f-8104-1c2c01c70d1b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.363119] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1179.363119] env[61911]: value = "task-1251737" [ 1179.363119] env[61911]: _type = "Task" [ 1179.363119] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.370235] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251737, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.408821] env[61911]: DEBUG nova.compute.manager [req-7e79cfb5-4ed8-489c-85d5-c198a3da85d8 req-f2045540-ecdd-4557-85fd-47e387230566 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-vif-plugged-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1179.409127] env[61911]: DEBUG oslo_concurrency.lockutils [req-7e79cfb5-4ed8-489c-85d5-c198a3da85d8 req-f2045540-ecdd-4557-85fd-47e387230566 service nova] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.409310] env[61911]: DEBUG oslo_concurrency.lockutils [req-7e79cfb5-4ed8-489c-85d5-c198a3da85d8 req-f2045540-ecdd-4557-85fd-47e387230566 service nova] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.409501] env[61911]: DEBUG oslo_concurrency.lockutils [req-7e79cfb5-4ed8-489c-85d5-c198a3da85d8 req-f2045540-ecdd-4557-85fd-47e387230566 service nova] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.409655] env[61911]: DEBUG nova.compute.manager [req-7e79cfb5-4ed8-489c-85d5-c198a3da85d8 req-f2045540-ecdd-4557-85fd-47e387230566 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] No waiting events found dispatching network-vif-plugged-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1179.409821] env[61911]: WARNING nova.compute.manager [req-7e79cfb5-4ed8-489c-85d5-c198a3da85d8 req-f2045540-ecdd-4557-85fd-47e387230566 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received unexpected event network-vif-plugged-cca0c90b-e103-4eca-9551-e89b65f5917a for instance with vm_state building and task_state spawning. [ 1179.496624] env[61911]: DEBUG nova.network.neutron [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Successfully updated port: cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1179.872455] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251737, 'name': Rename_Task, 'duration_secs': 0.13991} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.872832] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1179.873099] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7f9d47c-ae22-485e-875e-38098373856c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.878187] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1179.878187] env[61911]: value = "task-1251739" [ 1179.878187] env[61911]: _type = "Task" [ 1179.878187] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.886460] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251739, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.999887] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.000059] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.000246] env[61911]: DEBUG nova.network.neutron [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1180.388313] env[61911]: DEBUG oslo_vmware.api [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251739, 'name': PowerOnVM_Task, 'duration_secs': 0.40577} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.388549] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1180.388760] env[61911]: DEBUG nova.compute.manager [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1180.389512] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d57d20-9676-45ba-82d0-6e61b5d12d96 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.530397] env[61911]: DEBUG nova.network.neutron [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1180.663998] env[61911]: DEBUG nova.network.neutron [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.905102] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.905102] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.905102] env[61911]: DEBUG nova.objects.instance [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61911) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1181.167231] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.167511] env[61911]: DEBUG nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Instance network_info: |[{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1181.167971] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:93:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a58387dd-f438-4913-af6a-fafb734cd881', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cca0c90b-e103-4eca-9551-e89b65f5917a', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.175434] env[61911]: DEBUG oslo.service.loopingcall [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1181.175652] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1181.175881] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-639af883-477a-4e6d-8138-83e6d8effe2f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.189927] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "2077036f-3451-416c-8783-aa3ea5fbcbcb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.190179] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "2077036f-3451-416c-8783-aa3ea5fbcbcb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.190377] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "2077036f-3451-416c-8783-aa3ea5fbcbcb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.190564] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "2077036f-3451-416c-8783-aa3ea5fbcbcb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.190735] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "2077036f-3451-416c-8783-aa3ea5fbcbcb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.192759] env[61911]: INFO nova.compute.manager [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Terminating instance [ 1181.198951] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1181.198951] env[61911]: value = "task-1251740" [ 1181.198951] env[61911]: _type = "Task" [ 1181.198951] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.207437] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251740, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.437788] env[61911]: DEBUG nova.compute.manager [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1181.438019] env[61911]: DEBUG nova.compute.manager [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing instance network info cache due to event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1181.438260] env[61911]: DEBUG oslo_concurrency.lockutils [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.438400] env[61911]: DEBUG oslo_concurrency.lockutils [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.438593] env[61911]: DEBUG nova.network.neutron [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1181.696743] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "refresh_cache-2077036f-3451-416c-8783-aa3ea5fbcbcb" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.696937] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquired lock "refresh_cache-2077036f-3451-416c-8783-aa3ea5fbcbcb" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.697103] env[61911]: DEBUG nova.network.neutron [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1181.709353] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251740, 'name': CreateVM_Task, 'duration_secs': 0.379721} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.709553] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1181.710194] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.710435] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.710796] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1181.711334] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39917e3a-697c-4240-a3ee-94ee49ac2d9e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.717935] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1181.717935] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526e1303-5bbf-dc87-ec77-1ede35458551" [ 1181.717935] env[61911]: _type = "Task" [ 1181.717935] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.725294] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526e1303-5bbf-dc87-ec77-1ede35458551, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.914215] env[61911]: DEBUG oslo_concurrency.lockutils [None req-797f75bc-3cb9-457a-81c5-d60331588211 tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.143329] env[61911]: DEBUG nova.network.neutron [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updated VIF entry in instance network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1182.143710] env[61911]: DEBUG nova.network.neutron [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.214653] env[61911]: DEBUG nova.network.neutron [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1182.227661] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]526e1303-5bbf-dc87-ec77-1ede35458551, 'name': SearchDatastore_Task, 'duration_secs': 0.009604} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.229690] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.229929] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1182.230181] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.230336] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.230520] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.230810] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef7031fe-db18-4900-9810-c8482fb6b3f7 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.238479] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.238656] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1182.239361] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79e191a5-37cc-421f-99bb-966f169e37d2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.244078] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1182.244078] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5224bce1-8f58-edd0-a0ec-7aad11e9ac5c" [ 1182.244078] env[61911]: _type = "Task" [ 1182.244078] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.250848] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5224bce1-8f58-edd0-a0ec-7aad11e9ac5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.264499] env[61911]: DEBUG nova.network.neutron [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.646437] env[61911]: DEBUG oslo_concurrency.lockutils [req-e3fe76e2-b05f-467b-b77c-d9dafa198ce0 req-52c238e8-1e8d-411c-a611-d490b2e122ae service nova] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.754368] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5224bce1-8f58-edd0-a0ec-7aad11e9ac5c, 'name': SearchDatastore_Task, 'duration_secs': 0.00808} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.755160] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5944514-7820-4530-9be0-0cb7050cd4f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.760087] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1182.760087] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521f570a-8830-d05c-6798-d77162a168ef" [ 1182.760087] env[61911]: _type = "Task" [ 1182.760087] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.767945] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Releasing lock "refresh_cache-2077036f-3451-416c-8783-aa3ea5fbcbcb" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.768402] env[61911]: DEBUG nova.compute.manager [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1182.768617] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1182.768878] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521f570a-8830-d05c-6798-d77162a168ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.769578] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b93252-f291-4a07-835b-96d238cdba2d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.775851] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1182.776088] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15b6f135-4e32-4b49-ac7f-ef80726643a3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.781828] env[61911]: DEBUG oslo_vmware.api [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1182.781828] env[61911]: value = "task-1251741" [ 1182.781828] env[61911]: _type = "Task" [ 1182.781828] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.789725] env[61911]: DEBUG oslo_vmware.api [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251741, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.270370] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]521f570a-8830-d05c-6798-d77162a168ef, 'name': SearchDatastore_Task, 'duration_secs': 0.011597} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.270698] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.270909] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1183.271201] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8e62b5e-70ff-425c-a4d3-8c548303752d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.277358] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1183.277358] env[61911]: value = "task-1251743" [ 1183.277358] env[61911]: _type = "Task" [ 1183.277358] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.284381] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.290887] env[61911]: DEBUG oslo_vmware.api [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251741, 'name': PowerOffVM_Task, 'duration_secs': 0.114337} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.291171] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1183.291359] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1183.291581] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6dbc8a1f-6e40-43bd-aff0-d7db26e8c71b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.319415] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1183.319669] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1183.319856] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Deleting the datastore file [datastore1] 2077036f-3451-416c-8783-aa3ea5fbcbcb {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1183.320119] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66a0082b-c7c9-4d25-b5f3-0e2d79c95411 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.325868] env[61911]: DEBUG oslo_vmware.api [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for the task: (returnval){ [ 1183.325868] env[61911]: value = "task-1251745" [ 1183.325868] env[61911]: _type = "Task" [ 1183.325868] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.333075] env[61911]: DEBUG oslo_vmware.api [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.787822] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251743, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507513} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.788177] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1183.788337] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1183.788623] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3fbf3754-cc6d-49c1-8521-ec5236b0c202 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.794417] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1183.794417] env[61911]: value = "task-1251746" [ 1183.794417] env[61911]: _type = "Task" [ 1183.794417] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.802971] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.833696] env[61911]: DEBUG oslo_vmware.api [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Task: {'id': task-1251745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126316} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.833903] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1183.834088] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1183.834280] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1183.834464] env[61911]: INFO nova.compute.manager [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1183.834691] env[61911]: DEBUG oslo.service.loopingcall [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1183.834881] env[61911]: DEBUG nova.compute.manager [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1183.834976] env[61911]: DEBUG nova.network.neutron [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1183.849980] env[61911]: DEBUG nova.network.neutron [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1184.304183] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054674} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.304495] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1184.305253] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e374f2-5037-48e4-bbc1-d439f624c454 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.326294] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1184.326530] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67953179-e4ed-46fb-bc45-4155edf599ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.346017] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1184.346017] env[61911]: value = "task-1251747" [ 1184.346017] env[61911]: _type = "Task" [ 1184.346017] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.353197] env[61911]: DEBUG nova.network.neutron [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.354500] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.856303] env[61911]: INFO nova.compute.manager [-] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Took 1.02 seconds to deallocate network for instance. [ 1184.856540] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251747, 'name': ReconfigVM_Task, 'duration_secs': 0.249587} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.858277] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1184.861257] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f63a59a-10b0-4d53-8dbe-d0cb69afb698 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.868160] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1184.868160] env[61911]: value = "task-1251748" [ 1184.868160] env[61911]: _type = "Task" [ 1184.868160] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.875590] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251748, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.364422] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.364760] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.364994] env[61911]: DEBUG nova.objects.instance [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lazy-loading 'resources' on Instance uuid 2077036f-3451-416c-8783-aa3ea5fbcbcb {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1185.378054] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251748, 'name': Rename_Task, 'duration_secs': 0.143596} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.378300] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1185.378529] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b1be084-9527-4c4d-a9ea-ed1f91a19b02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.383602] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1185.383602] env[61911]: value = "task-1251749" [ 1185.383602] env[61911]: _type = "Task" [ 1185.383602] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.390895] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251749, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.894172] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251749, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.920203] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e69f94-9c81-435c-96c8-f2e739fb9344 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.927065] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014f4f38-ee5d-4103-adde-8a10d2eb0062 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.957329] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6fb085-94d4-47f0-b735-e7a462175c3d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.964336] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef56c7f-5903-4f8a-8d08-79abad278bda {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.977259] env[61911]: DEBUG nova.compute.provider_tree [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.394572] env[61911]: DEBUG oslo_vmware.api [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251749, 'name': PowerOnVM_Task, 'duration_secs': 0.528523} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.394866] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1186.395093] env[61911]: INFO nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Took 7.19 seconds to spawn the instance on the hypervisor. [ 1186.395287] env[61911]: DEBUG nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1186.396059] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1265130a-fb47-4477-9812-ee3d4bedc740 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.480443] env[61911]: DEBUG nova.scheduler.client.report [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1186.915355] env[61911]: INFO nova.compute.manager [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Took 12.27 seconds to build instance. [ 1186.984675] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.008442] env[61911]: INFO nova.scheduler.client.report [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Deleted allocations for instance 2077036f-3451-416c-8783-aa3ea5fbcbcb [ 1187.421097] env[61911]: DEBUG oslo_concurrency.lockutils [None req-4c63e34c-514c-48d2-bb8e-5d39302c365c tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.780s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.518337] env[61911]: DEBUG oslo_concurrency.lockutils [None req-723a2a36-14b0-4241-949d-a9a19e086ccd tempest-ServerShowV257Test-1029688195 tempest-ServerShowV257Test-1029688195-project-member] Lock "2077036f-3451-416c-8783-aa3ea5fbcbcb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.328s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.066892] env[61911]: DEBUG nova.compute.manager [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1188.067124] env[61911]: DEBUG nova.compute.manager [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing instance network info cache due to event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1188.067354] env[61911]: DEBUG oslo_concurrency.lockutils [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1188.067535] env[61911]: DEBUG oslo_concurrency.lockutils [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1188.067665] env[61911]: DEBUG nova.network.neutron [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1188.797825] env[61911]: DEBUG nova.network.neutron [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updated VIF entry in instance network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1188.798236] env[61911]: DEBUG nova.network.neutron [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.301864] env[61911]: DEBUG oslo_concurrency.lockutils [req-72e81cb2-ec90-4085-a9ae-a8167180f8d0 req-128c2294-d9e1-49b3-b521-bb9836cb4e98 service nova] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.370135] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.370430] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.872759] env[61911]: DEBUG nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1192.393757] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.394034] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.396077] env[61911]: INFO nova.compute.claims [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1193.448063] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abb480a-f69e-4399-ae57-4120ce1b2116 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.454632] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a6f0db-2497-48e0-9025-b5d8120b9096 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.485128] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe353d23-4b30-42a1-9a29-5fd502475142 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.492629] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6da349e-3d78-44d3-9fd2-f6764485308c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.505401] env[61911]: DEBUG nova.compute.provider_tree [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.008792] env[61911]: DEBUG nova.scheduler.client.report [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1194.512978] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.513571] env[61911]: DEBUG nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1195.019212] env[61911]: DEBUG nova.compute.utils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1195.020647] env[61911]: DEBUG nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1195.020820] env[61911]: DEBUG nova.network.neutron [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1195.059233] env[61911]: DEBUG nova.policy [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a874ae493bb84b848c196eb62c787bdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '104053ff00ae487692e030d91965a714', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1195.328419] env[61911]: DEBUG nova.network.neutron [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Successfully created port: fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1195.524929] env[61911]: DEBUG nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1196.537059] env[61911]: DEBUG nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1196.561416] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1196.561687] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1196.561851] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1196.562053] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1196.562233] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1196.562403] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1196.562612] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1196.562777] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1196.562951] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1196.563150] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1196.563366] env[61911]: DEBUG nova.virt.hardware [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1196.564259] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c21a225-9ff4-4dbf-a217-c91b9aad0712 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.571987] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60671b15-782a-4eb5-a1c2-9371909860a4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.712981] env[61911]: DEBUG nova.compute.manager [req-15dff8fb-79e1-42af-8cfc-4c49f1e682a3 req-a8123fad-dab9-486f-ab14-5392bfafd01c service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Received event network-vif-plugged-fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1196.713229] env[61911]: DEBUG oslo_concurrency.lockutils [req-15dff8fb-79e1-42af-8cfc-4c49f1e682a3 req-a8123fad-dab9-486f-ab14-5392bfafd01c service nova] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.713444] env[61911]: DEBUG oslo_concurrency.lockutils [req-15dff8fb-79e1-42af-8cfc-4c49f1e682a3 req-a8123fad-dab9-486f-ab14-5392bfafd01c service nova] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.713621] env[61911]: DEBUG oslo_concurrency.lockutils [req-15dff8fb-79e1-42af-8cfc-4c49f1e682a3 req-a8123fad-dab9-486f-ab14-5392bfafd01c service nova] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.713799] env[61911]: DEBUG nova.compute.manager [req-15dff8fb-79e1-42af-8cfc-4c49f1e682a3 req-a8123fad-dab9-486f-ab14-5392bfafd01c service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] No waiting events found dispatching network-vif-plugged-fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1196.713970] env[61911]: WARNING nova.compute.manager [req-15dff8fb-79e1-42af-8cfc-4c49f1e682a3 req-a8123fad-dab9-486f-ab14-5392bfafd01c service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Received unexpected event network-vif-plugged-fe976991-ceca-4699-ad13-b7768f1d9563 for instance with vm_state building and task_state spawning. [ 1196.794434] env[61911]: DEBUG nova.network.neutron [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Successfully updated port: fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1197.297735] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.297999] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.298113] env[61911]: DEBUG nova.network.neutron [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1197.829790] env[61911]: DEBUG nova.network.neutron [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1197.947076] env[61911]: DEBUG nova.network.neutron [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [{"id": "fe976991-ceca-4699-ad13-b7768f1d9563", "address": "fa:16:3e:23:89:e1", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe976991-ce", "ovs_interfaceid": "fe976991-ceca-4699-ad13-b7768f1d9563", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.450307] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.450638] env[61911]: DEBUG nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Instance network_info: |[{"id": "fe976991-ceca-4699-ad13-b7768f1d9563", "address": "fa:16:3e:23:89:e1", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe976991-ce", "ovs_interfaceid": "fe976991-ceca-4699-ad13-b7768f1d9563", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1198.451116] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:89:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1fb81f98-6f5a-47ab-a512-27277591d064', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe976991-ceca-4699-ad13-b7768f1d9563', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1198.458603] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Creating folder: Project (104053ff00ae487692e030d91965a714). Parent ref: group-v269521. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1198.458883] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3963481-31e4-4c7c-8da6-740f26b82fbc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.472747] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Created folder: Project (104053ff00ae487692e030d91965a714) in parent group-v269521. [ 1198.472902] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Creating folder: Instances. Parent ref: group-v269696. {{(pid=61911) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1198.473480] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8a875eb-5dac-43ea-b317-9f544474b922 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.480963] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Created folder: Instances in parent group-v269696. [ 1198.481223] env[61911]: DEBUG oslo.service.loopingcall [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.481408] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1198.481596] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d517471-af73-48cc-9b37-18bdcd678286 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.498488] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1198.498488] env[61911]: value = "task-1251752" [ 1198.498488] env[61911]: _type = "Task" [ 1198.498488] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.505211] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251752, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.740179] env[61911]: DEBUG nova.compute.manager [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Received event network-changed-fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1198.740434] env[61911]: DEBUG nova.compute.manager [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Refreshing instance network info cache due to event network-changed-fe976991-ceca-4699-ad13-b7768f1d9563. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1198.740567] env[61911]: DEBUG oslo_concurrency.lockutils [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] Acquiring lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1198.740759] env[61911]: DEBUG oslo_concurrency.lockutils [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] Acquired lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.740950] env[61911]: DEBUG nova.network.neutron [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Refreshing network info cache for port fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1199.008954] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251752, 'name': CreateVM_Task, 'duration_secs': 0.343072} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.009292] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1199.009794] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.009971] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.010320] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1199.010577] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc3bddf7-e89e-4d37-a44f-9862a266c341 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.014799] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1199.014799] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aee084-e680-5dc7-b277-1203fdfa071b" [ 1199.014799] env[61911]: _type = "Task" [ 1199.014799] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.021868] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aee084-e680-5dc7-b277-1203fdfa071b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.451635] env[61911]: DEBUG nova.network.neutron [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updated VIF entry in instance network info cache for port fe976991-ceca-4699-ad13-b7768f1d9563. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1199.451980] env[61911]: DEBUG nova.network.neutron [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [{"id": "fe976991-ceca-4699-ad13-b7768f1d9563", "address": "fa:16:3e:23:89:e1", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe976991-ce", "ovs_interfaceid": "fe976991-ceca-4699-ad13-b7768f1d9563", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.525216] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52aee084-e680-5dc7-b277-1203fdfa071b, 'name': SearchDatastore_Task, 'duration_secs': 0.008828} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.525761] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.526008] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1199.526258] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.526414] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.526604] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1199.526857] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c23c85f-34bc-4b2c-80ca-181926523fb8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.534301] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1199.534469] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1199.535149] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a262ce4-94d7-4fc9-b26a-958faef37133 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.540212] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1199.540212] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f7100f-5d93-ac59-6a80-8bcb03871704" [ 1199.540212] env[61911]: _type = "Task" [ 1199.540212] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.547151] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f7100f-5d93-ac59-6a80-8bcb03871704, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.954964] env[61911]: DEBUG oslo_concurrency.lockutils [req-dbf15575-1747-4eec-82ed-01e4219490fa req-e919cb3a-1748-4d5a-9509-8f5c050dbb40 service nova] Releasing lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.050373] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52f7100f-5d93-ac59-6a80-8bcb03871704, 'name': SearchDatastore_Task, 'duration_secs': 0.007951} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.051167] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70948dfb-a608-48ed-b178-2fbf63df9dc9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.055793] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1200.055793] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52db3a41-f0df-5259-2e75-83b3a10d19d7" [ 1200.055793] env[61911]: _type = "Task" [ 1200.055793] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.063651] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52db3a41-f0df-5259-2e75-83b3a10d19d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.566253] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52db3a41-f0df-5259-2e75-83b3a10d19d7, 'name': SearchDatastore_Task, 'duration_secs': 0.00866} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.566522] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.566791] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] bc36e378-9d5d-49a5-b7a6-8d4ca3303063/bc36e378-9d5d-49a5-b7a6-8d4ca3303063.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1200.567064] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a72fa3a3-da22-48d6-ad90-1875948b097b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.573085] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1200.573085] env[61911]: value = "task-1251753" [ 1200.573085] env[61911]: _type = "Task" [ 1200.573085] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.580332] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251753, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.083279] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251753, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.407084} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.083659] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] bc36e378-9d5d-49a5-b7a6-8d4ca3303063/bc36e378-9d5d-49a5-b7a6-8d4ca3303063.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1201.083734] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1201.083941] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be1f8672-37ee-4306-b910-32161511edda {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.089709] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1201.089709] env[61911]: value = "task-1251754" [ 1201.089709] env[61911]: _type = "Task" [ 1201.089709] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.096402] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251754, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.599518] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061778} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.599806] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1201.600572] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0414fd78-463a-4c6b-bf05-d4e0fdcc5c3b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.622861] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] bc36e378-9d5d-49a5-b7a6-8d4ca3303063/bc36e378-9d5d-49a5-b7a6-8d4ca3303063.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.622861] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40cfa85e-a11a-4920-9675-96402b306906 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.641438] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1201.641438] env[61911]: value = "task-1251755" [ 1201.641438] env[61911]: _type = "Task" [ 1201.641438] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.649111] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251755, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.151454] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251755, 'name': ReconfigVM_Task, 'duration_secs': 0.291275} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.151807] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Reconfigured VM instance instance-0000006d to attach disk [datastore1] bc36e378-9d5d-49a5-b7a6-8d4ca3303063/bc36e378-9d5d-49a5-b7a6-8d4ca3303063.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.152349] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73057622-6caf-45ab-b8e7-b97d074a35d0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.159087] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1202.159087] env[61911]: value = "task-1251756" [ 1202.159087] env[61911]: _type = "Task" [ 1202.159087] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.167086] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251756, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.668258] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251756, 'name': Rename_Task, 'duration_secs': 0.133473} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.668587] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1202.668791] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93468ae1-1ac7-4bac-8d99-9267471d5eb5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.675520] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1202.675520] env[61911]: value = "task-1251757" [ 1202.675520] env[61911]: _type = "Task" [ 1202.675520] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.682843] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251757, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.185711] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251757, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.685786] env[61911]: DEBUG oslo_vmware.api [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251757, 'name': PowerOnVM_Task, 'duration_secs': 0.562704} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.686165] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1203.686457] env[61911]: INFO nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1203.686717] env[61911]: DEBUG nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1203.687520] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0f3cad-c150-494d-bf47-79ec37e40b9b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.208888] env[61911]: INFO nova.compute.manager [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Took 11.83 seconds to build instance. [ 1204.376362] env[61911]: DEBUG nova.compute.manager [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Received event network-changed-fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1204.376569] env[61911]: DEBUG nova.compute.manager [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Refreshing instance network info cache due to event network-changed-fe976991-ceca-4699-ad13-b7768f1d9563. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1204.376793] env[61911]: DEBUG oslo_concurrency.lockutils [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] Acquiring lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1204.376941] env[61911]: DEBUG oslo_concurrency.lockutils [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] Acquired lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.377118] env[61911]: DEBUG nova.network.neutron [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Refreshing network info cache for port fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1204.711013] env[61911]: DEBUG oslo_concurrency.lockutils [None req-0cf69b73-ac0b-44b4-a58e-de07b491664c tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.340s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.079803] env[61911]: DEBUG nova.network.neutron [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updated VIF entry in instance network info cache for port fe976991-ceca-4699-ad13-b7768f1d9563. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1205.080198] env[61911]: DEBUG nova.network.neutron [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [{"id": "fe976991-ceca-4699-ad13-b7768f1d9563", "address": "fa:16:3e:23:89:e1", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe976991-ce", "ovs_interfaceid": "fe976991-ceca-4699-ad13-b7768f1d9563", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.582775] env[61911]: DEBUG oslo_concurrency.lockutils [req-7f708b8e-c83b-45ef-8ce8-9acdaf919f3d req-d08fa0d6-a10a-4fec-a4dd-a7c891e2594d service nova] Releasing lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1215.738105] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1216.241819] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.242105] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.242293] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.242453] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1216.243396] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923d4000-8d75-4b89-a37a-9a387818c4e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.251748] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b20e2a8-0bd9-4efe-93fd-fb2505fa11f9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.265413] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e61f8e-df3f-411b-847f-629cb5c5d628 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.271441] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b1a2c3-54f3-499c-8046-dcb217b3d03a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.300010] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181031MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1216.300162] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.300353] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.326536] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 90c0b35d-430a-4946-8a02-bae7f575e5ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1217.326786] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bc36e378-9d5d-49a5-b7a6-8d4ca3303063 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1217.326868] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1217.327055] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1217.362188] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5302f984-f784-4e4c-bb46-2092e0c58cb5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.369369] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f65e427-4305-4490-a842-41460dd3173e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.398344] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4b1862-446a-4b20-9e71-5035c1fb0e16 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.405323] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bd790b-0acd-4d05-a068-b4991e85498e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.417817] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.920636] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1218.425637] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1218.425996] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.125s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.420396] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.420632] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.420777] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.420927] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.421115] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.737274] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.737699] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1221.738454] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1221.738819] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1221.738819] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Rebuilding the list of instances to heal {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1222.269922] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.270271] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.270535] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1222.270810] env[61911]: DEBUG nova.objects.instance [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lazy-loading 'info_cache' on Instance uuid 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1223.994903] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.498336] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1224.498539] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1224.738028] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.243171] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.897744] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.897909] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.897951] env[61911]: INFO nova.compute.manager [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Shelving [ 1226.907594] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1226.907594] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca556bac-2db5-4835-82ed-2381982eb655 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.915999] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1226.915999] env[61911]: value = "task-1251758" [ 1226.915999] env[61911]: _type = "Task" [ 1226.915999] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.924676] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251758, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.425489] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251758, 'name': PowerOffVM_Task, 'duration_secs': 0.219368} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.425825] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1227.426633] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74fcabb-0ef7-4024-b52b-88a5d3497107 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.445039] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86639a5b-359b-40a0-beeb-3420de68c8ff {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.954627] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Creating Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1227.955063] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e8303a3d-5f64-4925-81fd-1da00345e45a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.963614] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1227.963614] env[61911]: value = "task-1251759" [ 1227.963614] env[61911]: _type = "Task" [ 1227.963614] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.972320] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251759, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.473751] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251759, 'name': CreateSnapshot_Task, 'duration_secs': 0.41048} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.474053] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Created Snapshot of the VM instance {{(pid=61911) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1228.474835] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ca936a-098f-4099-96dc-84460ea45257 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.991799] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Creating linked-clone VM from snapshot {{(pid=61911) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1228.992226] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-82f0a8c7-d428-4551-ad4c-ea8fc1241972 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.001163] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1229.001163] env[61911]: value = "task-1251760" [ 1229.001163] env[61911]: _type = "Task" [ 1229.001163] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.008641] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251760, 'name': CloneVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.511641] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251760, 'name': CloneVM_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.011714] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251760, 'name': CloneVM_Task, 'duration_secs': 0.989631} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.012111] env[61911]: INFO nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Created linked-clone VM from snapshot [ 1230.012716] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3357442-be89-4f1d-8d0c-1c63cfe902b2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.019562] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Uploading image 0ca2b7d4-6a8b-432d-9a1c-f6b938097837 {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1230.039403] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1230.039403] env[61911]: value = "vm-269700" [ 1230.039403] env[61911]: _type = "VirtualMachine" [ 1230.039403] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1230.039637] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8b6ac98e-e407-40b0-a572-4810a485394e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.045913] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease: (returnval){ [ 1230.045913] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d99ccb-4f17-e8cf-22af-41f3cf5b5c7f" [ 1230.045913] env[61911]: _type = "HttpNfcLease" [ 1230.045913] env[61911]: } obtained for exporting VM: (result){ [ 1230.045913] env[61911]: value = "vm-269700" [ 1230.045913] env[61911]: _type = "VirtualMachine" [ 1230.045913] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1230.046226] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the lease: (returnval){ [ 1230.046226] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d99ccb-4f17-e8cf-22af-41f3cf5b5c7f" [ 1230.046226] env[61911]: _type = "HttpNfcLease" [ 1230.046226] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1230.051897] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1230.051897] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d99ccb-4f17-e8cf-22af-41f3cf5b5c7f" [ 1230.051897] env[61911]: _type = "HttpNfcLease" [ 1230.051897] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1230.554978] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1230.554978] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d99ccb-4f17-e8cf-22af-41f3cf5b5c7f" [ 1230.554978] env[61911]: _type = "HttpNfcLease" [ 1230.554978] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1230.555318] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1230.555318] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d99ccb-4f17-e8cf-22af-41f3cf5b5c7f" [ 1230.555318] env[61911]: _type = "HttpNfcLease" [ 1230.555318] env[61911]: }. {{(pid=61911) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1230.556031] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94f195d-9523-43c7-8965-f2c3a215dd89 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.562965] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5268f361-ac5d-b119-b3c2-4171c93b00b2/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1230.563162] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5268f361-ac5d-b119-b3c2-4171c93b00b2/disk-0.vmdk for reading. {{(pid=61911) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1230.654255] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-58fdd7c4-e0a1-40c7-b031-c3a6f5794eb1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.963947] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5268f361-ac5d-b119-b3c2-4171c93b00b2/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1237.964644] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96afaa1d-f8d4-4989-a307-7b0d308cf7da {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.970672] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5268f361-ac5d-b119-b3c2-4171c93b00b2/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1237.970843] env[61911]: ERROR oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5268f361-ac5d-b119-b3c2-4171c93b00b2/disk-0.vmdk due to incomplete transfer. [ 1237.971067] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-43d2d023-ab6a-4648-8eb9-6d90c26ced0b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.986638] env[61911]: DEBUG oslo_vmware.rw_handles [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5268f361-ac5d-b119-b3c2-4171c93b00b2/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1237.986884] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Uploaded image 0ca2b7d4-6a8b-432d-9a1c-f6b938097837 to the Glance image server {{(pid=61911) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1237.989574] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Destroying the VM {{(pid=61911) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1237.989815] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f5387a3f-8bfd-47a3-a6e5-ce7038c65598 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.995411] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1237.995411] env[61911]: value = "task-1251762" [ 1237.995411] env[61911]: _type = "Task" [ 1237.995411] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.002753] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251762, 'name': Destroy_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.505610] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251762, 'name': Destroy_Task, 'duration_secs': 0.3182} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.505980] env[61911]: INFO nova.virt.vmwareapi.vm_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Destroyed the VM [ 1238.506322] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Deleting Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1238.506628] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-66a72133-7637-4925-8a76-a328e3f2edd8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.512777] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1238.512777] env[61911]: value = "task-1251763" [ 1238.512777] env[61911]: _type = "Task" [ 1238.512777] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.520264] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251763, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.023404] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251763, 'name': RemoveSnapshot_Task, 'duration_secs': 0.315495} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.023764] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Deleted Snapshot of the VM instance {{(pid=61911) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1239.024151] env[61911]: DEBUG nova.compute.manager [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1239.025048] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f178a1-4ae0-41c6-875f-1b96829f7dc8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.537170] env[61911]: INFO nova.compute.manager [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Shelve offloading [ 1240.040846] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1240.041279] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71a79300-d248-48ee-8a7b-66c9c5fe1fbe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.048669] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1240.048669] env[61911]: value = "task-1251764" [ 1240.048669] env[61911]: _type = "Task" [ 1240.048669] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.055955] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.558695] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] VM already powered off {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1240.558933] env[61911]: DEBUG nova.compute.manager [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1240.559716] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae714f9-5a10-4a3f-b8fa-932a17ea7910 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.565165] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1240.565352] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.565517] env[61911]: DEBUG nova.network.neutron [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1241.271473] env[61911]: DEBUG nova.network.neutron [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.774782] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1241.977738] env[61911]: DEBUG nova.compute.manager [req-f8f3dcd5-81e2-4105-b5b0-89e9afa50f87 req-8c572fc3-3264-40b2-ad55-5fad631b86d3 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-vif-unplugged-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1241.977943] env[61911]: DEBUG oslo_concurrency.lockutils [req-f8f3dcd5-81e2-4105-b5b0-89e9afa50f87 req-8c572fc3-3264-40b2-ad55-5fad631b86d3 service nova] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.978153] env[61911]: DEBUG oslo_concurrency.lockutils [req-f8f3dcd5-81e2-4105-b5b0-89e9afa50f87 req-8c572fc3-3264-40b2-ad55-5fad631b86d3 service nova] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.978331] env[61911]: DEBUG oslo_concurrency.lockutils [req-f8f3dcd5-81e2-4105-b5b0-89e9afa50f87 req-8c572fc3-3264-40b2-ad55-5fad631b86d3 service nova] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.978544] env[61911]: DEBUG nova.compute.manager [req-f8f3dcd5-81e2-4105-b5b0-89e9afa50f87 req-8c572fc3-3264-40b2-ad55-5fad631b86d3 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] No waiting events found dispatching network-vif-unplugged-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1241.978729] env[61911]: WARNING nova.compute.manager [req-f8f3dcd5-81e2-4105-b5b0-89e9afa50f87 req-8c572fc3-3264-40b2-ad55-5fad631b86d3 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received unexpected event network-vif-unplugged-cca0c90b-e103-4eca-9551-e89b65f5917a for instance with vm_state shelved and task_state shelving_offloading. [ 1242.080505] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1242.081451] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38324a34-bbd7-4e5f-9f47-bd3dd5e6f238 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.088492] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1242.088708] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34a1d900-6742-495f-9718-631c882d45ea {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.151021] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1242.151303] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1242.151500] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleting the datastore file [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1242.151767] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0bc670b4-7603-486b-841a-a3bc23d1bbad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.159067] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1242.159067] env[61911]: value = "task-1251766" [ 1242.159067] env[61911]: _type = "Task" [ 1242.159067] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.166112] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251766, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.670166] env[61911]: DEBUG oslo_vmware.api [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251766, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125455} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.670473] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1242.670668] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1242.670848] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1242.703120] env[61911]: INFO nova.scheduler.client.report [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted allocations for instance 90c0b35d-430a-4946-8a02-bae7f575e5ef [ 1243.208284] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.208557] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.208790] env[61911]: DEBUG nova.objects.instance [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'resources' on Instance uuid 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1243.709023] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.710024] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.711483] env[61911]: DEBUG nova.objects.instance [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'numa_topology' on Instance uuid 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.214546] env[61911]: DEBUG nova.compute.utils [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1244.215960] env[61911]: DEBUG nova.objects.base [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Object Instance<90c0b35d-430a-4946-8a02-bae7f575e5ef> lazy-loaded attributes: resources,numa_topology {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1244.243557] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc28b39d-c0d9-473a-938e-ff8380031f70 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.251283] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f008ec78-319f-484f-b22c-2554bd2cdfbb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.280486] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c0c83d-0ef1-4c3b-bbe0-365a09f53f84 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.287797] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b68a8ce-60e3-4c12-9ecc-e4a522d55de5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.302473] env[61911]: DEBUG nova.compute.provider_tree [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1244.418487] env[61911]: DEBUG nova.compute.manager [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1244.418585] env[61911]: DEBUG nova.compute.manager [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing instance network info cache due to event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1244.418794] env[61911]: DEBUG oslo_concurrency.lockutils [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.419108] env[61911]: DEBUG oslo_concurrency.lockutils [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.419334] env[61911]: DEBUG nova.network.neutron [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1244.721060] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.806221] env[61911]: DEBUG nova.scheduler.client.report [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1245.142840] env[61911]: DEBUG nova.network.neutron [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updated VIF entry in instance network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1245.143191] env[61911]: DEBUG nova.network.neutron [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.311405] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.614373] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1245.645282] env[61911]: DEBUG oslo_concurrency.lockutils [req-a142f0fe-5b11-46cd-860f-246c7b892148 req-45fff7cd-d19b-4c79-bcd6-d1470dd342c6 service nova] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.771564] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1245.771888] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.772093] env[61911]: INFO nova.compute.manager [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Attaching volume b93c0507-4aac-4b75-bfb9-f8b50f1e2c04 to /dev/sdb [ 1245.802837] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865bb4d9-7583-483a-86f3-843370bb40f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.809798] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5eb9a0b-2d9e-4ac6-9f86-691e1e5b7481 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.818680] env[61911]: DEBUG oslo_concurrency.lockutils [None req-41363ebb-ec7c-46d9-b798-20d9e3f59660 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.921s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.819917] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.205s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.819917] env[61911]: INFO nova.compute.manager [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Unshelving [ 1245.823853] env[61911]: DEBUG nova.virt.block_device [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating existing volume attachment record: 1afbce06-5a4c-4a77-ac96-e2c3463c27db {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1246.850084] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.850373] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.850565] env[61911]: DEBUG nova.objects.instance [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'pci_requests' on Instance uuid 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.353922] env[61911]: DEBUG nova.objects.instance [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'numa_topology' on Instance uuid 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.856745] env[61911]: INFO nova.compute.claims [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1248.902989] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf8525d-b4be-49e0-99c4-fbfa60a33592 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.910171] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4975c726-85a1-46d9-b127-dbf9638df6b1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.938626] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447aaaa2-773a-4234-92a3-3ab336c09248 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.945471] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d24f968-f68c-442a-9d89-35eda0c5affc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.957849] env[61911]: DEBUG nova.compute.provider_tree [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1249.460456] env[61911]: DEBUG nova.scheduler.client.report [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1249.965926] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.115s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.996377] env[61911]: INFO nova.network.neutron [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating port cca0c90b-e103-4eca-9551-e89b65f5917a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1250.369572] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1250.369827] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269702', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'name': 'volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc36e378-9d5d-49a5-b7a6-8d4ca3303063', 'attached_at': '', 'detached_at': '', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'serial': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1250.370698] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d4aaa8-b8b6-4757-957e-bd6a392319c5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.386571] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0b838d-718e-4816-8636-beb0ec60bbe1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.410355] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04/volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1250.410598] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bf21f82-97b2-4e30-8600-0aa9272adc61 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.427521] env[61911]: DEBUG oslo_vmware.api [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1250.427521] env[61911]: value = "task-1251771" [ 1250.427521] env[61911]: _type = "Task" [ 1250.427521] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.436363] env[61911]: DEBUG oslo_vmware.api [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251771, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.938285] env[61911]: DEBUG oslo_vmware.api [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251771, 'name': ReconfigVM_Task, 'duration_secs': 0.30995} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.938749] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04/volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1250.943604] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7ffb08e-3ef4-40eb-b883-d17b989ce347 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.958999] env[61911]: DEBUG oslo_vmware.api [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1250.958999] env[61911]: value = "task-1251772" [ 1250.958999] env[61911]: _type = "Task" [ 1250.958999] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.966531] env[61911]: DEBUG oslo_vmware.api [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251772, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.372618] env[61911]: DEBUG nova.compute.manager [req-6796f22c-99d2-4d98-b972-cddf7012e905 req-7ef61065-677a-4fe7-8985-17a04b1995d8 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-vif-plugged-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1251.372846] env[61911]: DEBUG oslo_concurrency.lockutils [req-6796f22c-99d2-4d98-b972-cddf7012e905 req-7ef61065-677a-4fe7-8985-17a04b1995d8 service nova] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.373073] env[61911]: DEBUG oslo_concurrency.lockutils [req-6796f22c-99d2-4d98-b972-cddf7012e905 req-7ef61065-677a-4fe7-8985-17a04b1995d8 service nova] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.373254] env[61911]: DEBUG oslo_concurrency.lockutils [req-6796f22c-99d2-4d98-b972-cddf7012e905 req-7ef61065-677a-4fe7-8985-17a04b1995d8 service nova] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.373499] env[61911]: DEBUG nova.compute.manager [req-6796f22c-99d2-4d98-b972-cddf7012e905 req-7ef61065-677a-4fe7-8985-17a04b1995d8 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] No waiting events found dispatching network-vif-plugged-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1251.373720] env[61911]: WARNING nova.compute.manager [req-6796f22c-99d2-4d98-b972-cddf7012e905 req-7ef61065-677a-4fe7-8985-17a04b1995d8 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received unexpected event network-vif-plugged-cca0c90b-e103-4eca-9551-e89b65f5917a for instance with vm_state shelved_offloaded and task_state spawning. [ 1251.449274] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.449505] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.449662] env[61911]: DEBUG nova.network.neutron [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1251.469902] env[61911]: DEBUG oslo_vmware.api [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251772, 'name': ReconfigVM_Task, 'duration_secs': 0.127112} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.470039] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269702', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'name': 'volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc36e378-9d5d-49a5-b7a6-8d4ca3303063', 'attached_at': '', 'detached_at': '', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'serial': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1252.154897] env[61911]: DEBUG nova.network.neutron [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.507634] env[61911]: DEBUG nova.objects.instance [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1252.657702] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1252.681912] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f9d7e995af754f29ad714f3f2169b86f',container_format='bare',created_at=2024-10-10T16:05:39Z,direct_url=,disk_format='vmdk',id=0ca2b7d4-6a8b-432d-9a1c-f6b938097837,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1740625555-shelved',owner='ce8675694c2841a58e87bb7250f3435a',properties=ImageMetaProps,protected=,size=31670784,status='active',tags=,updated_at=2024-10-10T16:05:51Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1252.682172] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1252.682359] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1252.682557] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1252.682710] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1252.682863] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1252.683084] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1252.683253] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1252.683425] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1252.683598] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1252.683797] env[61911]: DEBUG nova.virt.hardware [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1252.684673] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e7c49f-72a6-4ea4-9e62-384c56b9ab3d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.692314] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d21926-1a06-4f96-a6c9-995b6bf06a9a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.705707] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:93:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a58387dd-f438-4913-af6a-fafb734cd881', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cca0c90b-e103-4eca-9551-e89b65f5917a', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1252.712361] env[61911]: DEBUG oslo.service.loopingcall [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1252.712666] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1252.712940] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-988b694a-8c77-4af8-ad50-e0f4faf01cad {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.731947] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1252.731947] env[61911]: value = "task-1251773" [ 1252.731947] env[61911]: _type = "Task" [ 1252.731947] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.738664] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251773, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.012777] env[61911]: DEBUG oslo_concurrency.lockutils [None req-8c4941d9-13ff-4982-b0fc-3bce61782da0 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.085901] env[61911]: DEBUG oslo_concurrency.lockutils [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.086182] env[61911]: DEBUG oslo_concurrency.lockutils [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.086403] env[61911]: DEBUG nova.compute.manager [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1253.087329] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc55acaf-4331-489c-8672-a74808e0fc38 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.094027] env[61911]: DEBUG nova.compute.manager [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61911) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1253.094577] env[61911]: DEBUG nova.objects.instance [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.241466] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251773, 'name': CreateVM_Task, 'duration_secs': 0.317279} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.241908] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1253.242317] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1253.242524] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.242902] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1253.243172] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bde8ca94-863d-49e6-8489-976d0d52e6cd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.247773] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1253.247773] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b32a53-6d48-6009-1bfb-bf482fde27ae" [ 1253.247773] env[61911]: _type = "Task" [ 1253.247773] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.255655] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52b32a53-6d48-6009-1bfb-bf482fde27ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.396307] env[61911]: DEBUG nova.compute.manager [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1253.396503] env[61911]: DEBUG nova.compute.manager [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing instance network info cache due to event network-changed-cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1253.396727] env[61911]: DEBUG oslo_concurrency.lockutils [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] Acquiring lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1253.396879] env[61911]: DEBUG oslo_concurrency.lockutils [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] Acquired lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.397056] env[61911]: DEBUG nova.network.neutron [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Refreshing network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1253.761995] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1253.762450] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Processing image 0ca2b7d4-6a8b-432d-9a1c-f6b938097837 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1253.762843] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837/0ca2b7d4-6a8b-432d-9a1c-f6b938097837.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1253.763128] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837/0ca2b7d4-6a8b-432d-9a1c-f6b938097837.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.763450] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1253.763785] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c0001cc-e17e-4014-b054-8143c162f618 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.773169] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1253.773456] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1253.774501] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b963f1f-4927-49c8-af9c-a23a0bd06799 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.780705] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1253.780705] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e5c323-8fb3-0087-6253-6cee9dcc1638" [ 1253.780705] env[61911]: _type = "Task" [ 1253.780705] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.793521] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e5c323-8fb3-0087-6253-6cee9dcc1638, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.092075] env[61911]: DEBUG nova.network.neutron [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updated VIF entry in instance network info cache for port cca0c90b-e103-4eca-9551-e89b65f5917a. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1254.092488] env[61911]: DEBUG nova.network.neutron [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [{"id": "cca0c90b-e103-4eca-9551-e89b65f5917a", "address": "fa:16:3e:d2:93:26", "network": {"id": "cd70dfa4-3544-4daa-bc05-4b7d20819a54", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-419023781-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.219", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ce8675694c2841a58e87bb7250f3435a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca0c90b-e1", "ovs_interfaceid": "cca0c90b-e103-4eca-9551-e89b65f5917a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1254.100637] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1254.100913] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8f0da30-c208-4904-abe6-af8bd8c70504 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.108183] env[61911]: DEBUG oslo_vmware.api [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1254.108183] env[61911]: value = "task-1251774" [ 1254.108183] env[61911]: _type = "Task" [ 1254.108183] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.116013] env[61911]: DEBUG oslo_vmware.api [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251774, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.291130] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Preparing fetch location {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1254.291569] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Fetch image to [datastore1] OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5/OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5.vmdk {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1254.291814] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Downloading stream optimized image 0ca2b7d4-6a8b-432d-9a1c-f6b938097837 to [datastore1] OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5/OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5.vmdk on the data store datastore1 as vApp {{(pid=61911) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1254.292096] env[61911]: DEBUG nova.virt.vmwareapi.images [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Downloading image file data 0ca2b7d4-6a8b-432d-9a1c-f6b938097837 to the ESX as VM named 'OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5' {{(pid=61911) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1254.359286] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1254.359286] env[61911]: value = "resgroup-9" [ 1254.359286] env[61911]: _type = "ResourcePool" [ 1254.359286] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1254.359588] env[61911]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5a003a47-1fa6-44e8-b152-fd762b783446 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.380013] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease: (returnval){ [ 1254.380013] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5205c005-0ded-c95f-2598-ed88fc1d2336" [ 1254.380013] env[61911]: _type = "HttpNfcLease" [ 1254.380013] env[61911]: } obtained for vApp import into resource pool (val){ [ 1254.380013] env[61911]: value = "resgroup-9" [ 1254.380013] env[61911]: _type = "ResourcePool" [ 1254.380013] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1254.380287] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the lease: (returnval){ [ 1254.380287] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5205c005-0ded-c95f-2598-ed88fc1d2336" [ 1254.380287] env[61911]: _type = "HttpNfcLease" [ 1254.380287] env[61911]: } to be ready. {{(pid=61911) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1254.386111] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1254.386111] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5205c005-0ded-c95f-2598-ed88fc1d2336" [ 1254.386111] env[61911]: _type = "HttpNfcLease" [ 1254.386111] env[61911]: } is initializing. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1254.595770] env[61911]: DEBUG oslo_concurrency.lockutils [req-76a444d7-7faf-4989-b7d6-f42b00f528cc req-372458b5-808b-4772-a824-4d6243575215 service nova] Releasing lock "refresh_cache-90c0b35d-430a-4946-8a02-bae7f575e5ef" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1254.618623] env[61911]: DEBUG oslo_vmware.api [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251774, 'name': PowerOffVM_Task, 'duration_secs': 0.187747} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.618920] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1254.619173] env[61911]: DEBUG nova.compute.manager [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1254.619957] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380f229e-7e29-4853-bee8-92c0d62b6f65 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.888187] env[61911]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1254.888187] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5205c005-0ded-c95f-2598-ed88fc1d2336" [ 1254.888187] env[61911]: _type = "HttpNfcLease" [ 1254.888187] env[61911]: } is ready. {{(pid=61911) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1254.888429] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1254.888429] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]5205c005-0ded-c95f-2598-ed88fc1d2336" [ 1254.888429] env[61911]: _type = "HttpNfcLease" [ 1254.888429] env[61911]: }. {{(pid=61911) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1254.889142] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df78e7e-c5d7-49fe-8d86-e2202310a3a0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.895946] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52378dd9-2baf-7f49-ff43-1d74c69c39bb/disk-0.vmdk from lease info. {{(pid=61911) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1254.896144] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating HTTP connection to write to file with size = 31670784 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52378dd9-2baf-7f49-ff43-1d74c69c39bb/disk-0.vmdk. {{(pid=61911) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1254.959347] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5f60eb2b-b54d-4b77-a94b-f1e9179df7f8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.130950] env[61911]: DEBUG oslo_concurrency.lockutils [None req-efbc3009-e0d2-461a-bd90-b2fb83fbfaf9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.045s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.593549] env[61911]: DEBUG nova.objects.instance [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.099111] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.099358] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.099412] env[61911]: DEBUG nova.network.neutron [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1256.099584] env[61911]: DEBUG nova.objects.instance [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'info_cache' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.139438] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Completed reading data from the image iterator. {{(pid=61911) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1256.139701] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52378dd9-2baf-7f49-ff43-1d74c69c39bb/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1256.140667] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c13b03ad-82aa-4214-8109-dd5903a3d538 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.148049] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52378dd9-2baf-7f49-ff43-1d74c69c39bb/disk-0.vmdk is in state: ready. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1256.148246] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52378dd9-2baf-7f49-ff43-1d74c69c39bb/disk-0.vmdk. {{(pid=61911) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1256.148506] env[61911]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-037aab67-456d-4766-a6c6-b8c4fa92cb02 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.379923] env[61911]: DEBUG oslo_vmware.rw_handles [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52378dd9-2baf-7f49-ff43-1d74c69c39bb/disk-0.vmdk. {{(pid=61911) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1256.380097] env[61911]: INFO nova.virt.vmwareapi.images [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Downloaded image file data 0ca2b7d4-6a8b-432d-9a1c-f6b938097837 [ 1256.381056] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb90898-e513-4abd-b9f7-bb28dccbdad0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.399663] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ce4da98-cd79-4fc1-be3b-92c010819b0a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.420058] env[61911]: INFO nova.virt.vmwareapi.images [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] The imported VM was unregistered [ 1256.422530] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Caching image {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1256.422758] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Creating directory with path [datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1256.423025] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32201fd9-3813-4992-b6fe-99d816e173a1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.507503] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Created directory with path [datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837 {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1256.507690] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5/OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5.vmdk to [datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837/0ca2b7d4-6a8b-432d-9a1c-f6b938097837.vmdk. {{(pid=61911) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1256.507952] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e8c1f45f-269b-456a-9d59-ca93910625fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.515010] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1256.515010] env[61911]: value = "task-1251777" [ 1256.515010] env[61911]: _type = "Task" [ 1256.515010] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.522241] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251777, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.602812] env[61911]: DEBUG nova.objects.base [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1257.026616] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251777, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.478666] env[61911]: DEBUG nova.network.neutron [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [{"id": "fe976991-ceca-4699-ad13-b7768f1d9563", "address": "fa:16:3e:23:89:e1", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe976991-ce", "ovs_interfaceid": "fe976991-ceca-4699-ad13-b7768f1d9563", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.528469] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251777, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.981759] env[61911]: DEBUG oslo_concurrency.lockutils [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.029566] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251777, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.530899] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251777, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.988339] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1258.988766] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33222dad-8e36-44ec-b9b3-2da6421b4d8b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.996198] env[61911]: DEBUG oslo_vmware.api [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1258.996198] env[61911]: value = "task-1251778" [ 1258.996198] env[61911]: _type = "Task" [ 1258.996198] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.003981] env[61911]: DEBUG oslo_vmware.api [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251778, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.027438] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251777, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.158144} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.027714] env[61911]: INFO nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5/OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5.vmdk to [datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837/0ca2b7d4-6a8b-432d-9a1c-f6b938097837.vmdk. [ 1259.027883] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Cleaning up location [datastore1] OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5 {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1259.028081] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_cb133c05-cabb-4085-ab36-07857eb3b6d5 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1259.028364] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49c8779b-7930-4f83-af62-532c882cc7de {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.033469] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1259.033469] env[61911]: value = "task-1251779" [ 1259.033469] env[61911]: _type = "Task" [ 1259.033469] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.040133] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.506180] env[61911]: DEBUG oslo_vmware.api [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251778, 'name': PowerOnVM_Task, 'duration_secs': 0.404398} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.506462] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1259.506674] env[61911]: DEBUG nova.compute.manager [None req-d00b2fc6-e410-4fab-9590-fdf771845a4b tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1259.507431] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f60a3c-dcd1-4fa5-8ffe-1e339612d011 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.541696] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082494} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.541942] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1259.542130] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837/0ca2b7d4-6a8b-432d-9a1c-f6b938097837.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.542404] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837/0ca2b7d4-6a8b-432d-9a1c-f6b938097837.vmdk to [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1259.543276] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-367c052e-3def-4217-b1eb-c8bf893d7038 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.549174] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1259.549174] env[61911]: value = "task-1251780" [ 1259.549174] env[61911]: _type = "Task" [ 1259.549174] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.557550] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251780, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.060908] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251780, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.562406] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251780, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.064292] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251780, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.563747] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251780, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.063127] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251780, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.249082} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.063380] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0ca2b7d4-6a8b-432d-9a1c-f6b938097837/0ca2b7d4-6a8b-432d-9a1c-f6b938097837.vmdk to [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1262.064139] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a968a2-3202-4b4b-995f-c96314c45914 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.087505] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1262.087914] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfdda5b2-e490-4e39-8074-4dae7315e95c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.106980] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1262.106980] env[61911]: value = "task-1251781" [ 1262.106980] env[61911]: _type = "Task" [ 1262.106980] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.114493] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251781, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.616850] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251781, 'name': ReconfigVM_Task, 'duration_secs': 0.290487} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.617153] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef/90c0b35d-430a-4946-8a02-bae7f575e5ef.vmdk or device None with type streamOptimized {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1262.617773] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3a614fd-4ce8-4c74-8539-883a162c08cd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.623477] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1262.623477] env[61911]: value = "task-1251782" [ 1262.623477] env[61911]: _type = "Task" [ 1262.623477] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.630322] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251782, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.133659] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251782, 'name': Rename_Task, 'duration_secs': 0.169154} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.134039] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1263.134191] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15a24718-f876-4363-af19-b7d770527eb2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.140757] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1263.140757] env[61911]: value = "task-1251783" [ 1263.140757] env[61911]: _type = "Task" [ 1263.140757] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.148071] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.650995] env[61911]: DEBUG oslo_vmware.api [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251783, 'name': PowerOnVM_Task, 'duration_secs': 0.434235} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.651283] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1263.744243] env[61911]: DEBUG nova.compute.manager [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1263.745162] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9baf8561-703f-4fdf-8615-2b942e89c18c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.261670] env[61911]: DEBUG oslo_concurrency.lockutils [None req-c321afe0-1315-4cee-bc2a-b8c7ff40bb17 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.442s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.737995] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1276.241343] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.241642] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1276.241813] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1276.241976] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1276.242925] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f772df07-07cb-4960-aa19-e30077207178 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.251376] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38e77b0-c0ce-4b4c-9eb7-6a966573486b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.264923] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20fb263-1136-47ee-8801-3e4c049c9f18 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.271019] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9365db-ad29-4d37-9f74-7b49ff1dc657 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.298549] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181464MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1276.298708] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1276.298896] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.327281] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bc36e378-9d5d-49a5-b7a6-8d4ca3303063 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1277.327663] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance 90c0b35d-430a-4946-8a02-bae7f575e5ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1277.327663] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1277.327808] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1277.367812] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32605394-2eb6-4378-851c-85129b1d7c8a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.375216] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0cce81-d9c3-40f8-9f53-9d60f09a39f6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.067712] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1358b79f-2120-4026-89d6-efd578d1bf16 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.074861] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d0767e-3e59-481b-8f51-205c54b2c4d5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.087158] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1278.590694] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1279.095660] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1279.095911] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.797s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1281.090532] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.090962] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.091122] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.091289] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.091449] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.737128] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.737525] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1283.738813] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.739166] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1284.270234] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1284.270389] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.270557] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1285.489549] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [{"id": "fe976991-ceca-4699-ad13-b7768f1d9563", "address": "fa:16:3e:23:89:e1", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe976991-ce", "ovs_interfaceid": "fe976991-ceca-4699-ad13-b7768f1d9563", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1285.992327] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.992637] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1285.992883] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1295.840898] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.841303] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.344960] env[61911]: INFO nova.compute.manager [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Detaching volume b93c0507-4aac-4b75-bfb9-f8b50f1e2c04 [ 1296.377022] env[61911]: INFO nova.virt.block_device [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Attempting to driver detach volume b93c0507-4aac-4b75-bfb9-f8b50f1e2c04 from mountpoint /dev/sdb [ 1296.377022] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1296.377022] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269702', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'name': 'volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc36e378-9d5d-49a5-b7a6-8d4ca3303063', 'attached_at': '', 'detached_at': '', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'serial': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1296.377022] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85ab1bd-083e-4898-ba6a-b97076c9ca67 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.400727] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c8f294-ce93-4288-90ca-078bf142a806 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.407750] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1576bef-047e-41ae-8b44-999c2701c119 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.429229] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d65065-285f-4f10-a704-958f0375b3eb {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.445473] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] The volume has not been displaced from its original location: [datastore1] volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04/volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1296.452061] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1296.452061] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d089cf1-a3eb-4334-b296-49380f2c597c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.469932] env[61911]: DEBUG oslo_vmware.api [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1296.469932] env[61911]: value = "task-1251784" [ 1296.469932] env[61911]: _type = "Task" [ 1296.469932] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1296.477669] env[61911]: DEBUG oslo_vmware.api [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251784, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.979769] env[61911]: DEBUG oslo_vmware.api [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251784, 'name': ReconfigVM_Task, 'duration_secs': 0.218891} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.980213] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1296.985330] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7a0d9b5-e568-4900-9bd3-46e977cd9b97 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.000923] env[61911]: DEBUG oslo_vmware.api [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1297.000923] env[61911]: value = "task-1251785" [ 1297.000923] env[61911]: _type = "Task" [ 1297.000923] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.008390] env[61911]: DEBUG oslo_vmware.api [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251785, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.510599] env[61911]: DEBUG oslo_vmware.api [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251785, 'name': ReconfigVM_Task, 'duration_secs': 0.173714} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1297.510912] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269702', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'name': 'volume-b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bc36e378-9d5d-49a5-b7a6-8d4ca3303063', 'attached_at': '', 'detached_at': '', 'volume_id': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04', 'serial': 'b93c0507-4aac-4b75-bfb9-f8b50f1e2c04'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1298.051587] env[61911]: DEBUG nova.objects.instance [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1299.008969] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1299.059437] env[61911]: DEBUG oslo_concurrency.lockutils [None req-6519fdb7-9d9f-412f-9668-16eacec49cb9 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.218s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1299.060688] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.052s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.061057] env[61911]: DEBUG nova.compute.manager [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1299.062306] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a8c67f-6427-4c0b-beaa-0a227f364763 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.069359] env[61911]: DEBUG nova.compute.manager [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61911) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1299.069918] env[61911]: DEBUG nova.objects.instance [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1299.736763] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1299.737083] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.737686] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1299.737897] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.738094] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1299.740288] env[61911]: INFO nova.compute.manager [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Terminating instance [ 1300.076608] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1300.077077] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc033546-715a-48a1-aaa7-a8525761f117 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.084685] env[61911]: DEBUG oslo_vmware.api [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1300.084685] env[61911]: value = "task-1251786" [ 1300.084685] env[61911]: _type = "Task" [ 1300.084685] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.092661] env[61911]: DEBUG oslo_vmware.api [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251786, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.243877] env[61911]: DEBUG nova.compute.manager [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1300.244176] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1300.245248] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555eebf0-e13e-436a-a362-0662bd66ade0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.252648] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1300.252896] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7af941ab-0f27-4836-96bb-3f0d0b664f40 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.258753] env[61911]: DEBUG oslo_vmware.api [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1300.258753] env[61911]: value = "task-1251787" [ 1300.258753] env[61911]: _type = "Task" [ 1300.258753] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.266565] env[61911]: DEBUG oslo_vmware.api [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.594224] env[61911]: DEBUG oslo_vmware.api [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251786, 'name': PowerOffVM_Task, 'duration_secs': 0.166978} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.594494] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1300.594690] env[61911]: DEBUG nova.compute.manager [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1300.595425] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac1d91a-3845-4787-af38-491f02e9bc7f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.768762] env[61911]: DEBUG oslo_vmware.api [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251787, 'name': PowerOffVM_Task, 'duration_secs': 0.195011} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.769128] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1300.769319] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1300.769573] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7690b63b-b31d-409f-944b-78a3168047d0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.832808] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1300.833231] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1300.833547] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleting the datastore file [datastore1] 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1300.833939] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0833ac23-2b00-4925-aa95-1db1c2e50a5e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.841751] env[61911]: DEBUG oslo_vmware.api [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for the task: (returnval){ [ 1300.841751] env[61911]: value = "task-1251789" [ 1300.841751] env[61911]: _type = "Task" [ 1300.841751] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.853125] env[61911]: DEBUG oslo_vmware.api [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251789, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.106563] env[61911]: DEBUG oslo_concurrency.lockutils [None req-b27f71bb-02bf-43c3-8716-29191e609f0d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.046s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.351587] env[61911]: DEBUG oslo_vmware.api [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Task: {'id': task-1251789, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123427} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.351866] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1301.352085] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1301.352270] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1301.352449] env[61911]: INFO nova.compute.manager [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1301.352693] env[61911]: DEBUG oslo.service.loopingcall [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1301.352927] env[61911]: DEBUG nova.compute.manager [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1301.353038] env[61911]: DEBUG nova.network.neutron [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1301.395369] env[61911]: DEBUG nova.objects.instance [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1301.825840] env[61911]: DEBUG nova.compute.manager [req-30b3daca-bbaf-4035-9a2e-c0b16812793f req-2b9f71ae-6837-4a6c-97d6-bba3595baffb service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Received event network-vif-deleted-cca0c90b-e103-4eca-9551-e89b65f5917a {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1301.826058] env[61911]: INFO nova.compute.manager [req-30b3daca-bbaf-4035-9a2e-c0b16812793f req-2b9f71ae-6837-4a6c-97d6-bba3595baffb service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Neutron deleted interface cca0c90b-e103-4eca-9551-e89b65f5917a; detaching it from the instance and deleting it from the info cache [ 1301.826200] env[61911]: DEBUG nova.network.neutron [req-30b3daca-bbaf-4035-9a2e-c0b16812793f req-2b9f71ae-6837-4a6c-97d6-bba3595baffb service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1301.900108] env[61911]: DEBUG oslo_concurrency.lockutils [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1301.900286] env[61911]: DEBUG oslo_concurrency.lockutils [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.900462] env[61911]: DEBUG nova.network.neutron [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1301.900680] env[61911]: DEBUG nova.objects.instance [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'info_cache' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1302.306157] env[61911]: DEBUG nova.network.neutron [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.328675] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-728c30c5-a221-4f35-9e0f-d9661bc5f727 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.337564] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1abea82-3094-46fa-b540-fff73767b21f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.359890] env[61911]: DEBUG nova.compute.manager [req-30b3daca-bbaf-4035-9a2e-c0b16812793f req-2b9f71ae-6837-4a6c-97d6-bba3595baffb service nova] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Detach interface failed, port_id=cca0c90b-e103-4eca-9551-e89b65f5917a, reason: Instance 90c0b35d-430a-4946-8a02-bae7f575e5ef could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1302.404919] env[61911]: DEBUG nova.objects.base [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61911) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1302.809112] env[61911]: INFO nova.compute.manager [-] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Took 1.46 seconds to deallocate network for instance. [ 1303.118144] env[61911]: DEBUG nova.network.neutron [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [{"id": "fe976991-ceca-4699-ad13-b7768f1d9563", "address": "fa:16:3e:23:89:e1", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe976991-ce", "ovs_interfaceid": "fe976991-ceca-4699-ad13-b7768f1d9563", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1303.315898] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1303.316292] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1303.316544] env[61911]: DEBUG nova.objects.instance [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lazy-loading 'resources' on Instance uuid 90c0b35d-430a-4946-8a02-bae7f575e5ef {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1303.621389] env[61911]: DEBUG oslo_concurrency.lockutils [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1303.857642] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19176543-675d-4563-a122-d5f1d1bdc03a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.865050] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dc0ef8-1c0b-491d-b2c8-c51e174e1644 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.894953] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e54e0c-8f87-4dd2-99a3-002ef81c258a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.902048] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07e07d3-4ba8-485f-939e-98ac2f08e6e3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.914752] env[61911]: DEBUG nova.compute.provider_tree [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1304.418114] env[61911]: DEBUG nova.scheduler.client.report [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1304.627532] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1304.627830] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6b6817f-800b-4f1e-91a3-a93197dfd4b4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.635423] env[61911]: DEBUG oslo_vmware.api [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1304.635423] env[61911]: value = "task-1251790" [ 1304.635423] env[61911]: _type = "Task" [ 1304.635423] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.643116] env[61911]: DEBUG oslo_vmware.api [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251790, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.923258] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1304.943870] env[61911]: INFO nova.scheduler.client.report [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Deleted allocations for instance 90c0b35d-430a-4946-8a02-bae7f575e5ef [ 1305.144852] env[61911]: DEBUG oslo_vmware.api [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251790, 'name': PowerOnVM_Task, 'duration_secs': 0.34963} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.145151] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1305.145338] env[61911]: DEBUG nova.compute.manager [None req-55554a19-5d2f-4e1d-bc80-c90d130cc887 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1305.146110] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19079db5-d850-461c-b213-8ad1b9a11bf6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.453903] env[61911]: DEBUG oslo_concurrency.lockutils [None req-952adc6c-bae1-420c-88ea-b46a558a27d9 tempest-AttachVolumeShelveTestJSON-1650903832 tempest-AttachVolumeShelveTestJSON-1650903832-project-member] Lock "90c0b35d-430a-4946-8a02-bae7f575e5ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.717s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1337.738550] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.738952] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1337.738996] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1338.242740] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.243008] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1338.243212] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1338.243392] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1338.244305] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc0409d-8e78-439e-9df0-49fbb4de2d6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.252209] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162b999b-c167-4824-b993-8c9d10c9ee13 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.265606] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62412e67-711b-4e1d-8742-13c30d9ac2ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.271427] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9ed314-900e-400f-a445-2dbcb5779182 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.299272] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181302MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1338.299412] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.299645] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.325259] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance bc36e378-9d5d-49a5-b7a6-8d4ca3303063 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1339.325531] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1339.325633] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1339.341326] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Refreshing inventories for resource provider b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1339.354054] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Updating ProviderTree inventory for provider b8a56394-897a-4e67-92ba-db46db6115e9 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1339.354243] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Updating inventory in ProviderTree for provider b8a56394-897a-4e67-92ba-db46db6115e9 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1339.364226] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Refreshing aggregate associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, aggregates: None {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1339.381771] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Refreshing trait associations for resource provider b8a56394-897a-4e67-92ba-db46db6115e9, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61911) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1339.406261] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c393d060-2bd5-48ce-9791-7d99431c6dee {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.412990] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eaed6ad-bdec-40a8-afe1-f90769cc7735 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.442281] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed58b31-6471-43c5-bba8-9758aac70430 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.448647] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02fdf0b-d9d7-460d-9b72-fd964b9da669 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.460844] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1339.964221] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1340.469346] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1340.469716] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.170s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1341.461243] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1341.461598] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1341.461890] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1341.462137] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1341.462400] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1341.465775] env[61911]: INFO nova.compute.manager [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Terminating instance [ 1341.969990] env[61911]: DEBUG nova.compute.manager [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1341.970388] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1341.971166] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fde2f7-e212-4bf0-9ee5-c1244c3d85f1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.978936] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1341.979180] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ccb2dab-51be-4bff-9138-081e2b8a3ef0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.985201] env[61911]: DEBUG oslo_vmware.api [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1341.985201] env[61911]: value = "task-1251792" [ 1341.985201] env[61911]: _type = "Task" [ 1341.985201] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.991898] env[61911]: DEBUG oslo_vmware.api [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.495129] env[61911]: DEBUG oslo_vmware.api [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251792, 'name': PowerOffVM_Task, 'duration_secs': 0.154768} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.495399] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1342.495574] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1342.495815] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2d8845f-df8e-40b7-9da5-9effb8e46984 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.583195] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1342.583417] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1342.583578] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Deleting the datastore file [datastore1] bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1342.583848] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f1e6e70-8af6-4e01-bffe-48e3d43d4788 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.591153] env[61911]: DEBUG oslo_vmware.api [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1342.591153] env[61911]: value = "task-1251794" [ 1342.591153] env[61911]: _type = "Task" [ 1342.591153] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.598085] env[61911]: DEBUG oslo_vmware.api [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251794, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.100993] env[61911]: DEBUG oslo_vmware.api [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251794, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13536} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.101389] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1343.101499] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1343.101633] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1343.101813] env[61911]: INFO nova.compute.manager [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1343.102071] env[61911]: DEBUG oslo.service.loopingcall [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1343.102265] env[61911]: DEBUG nova.compute.manager [-] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1343.102378] env[61911]: DEBUG nova.network.neutron [-] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1343.468801] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1343.469070] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1343.469243] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1343.562952] env[61911]: DEBUG nova.compute.manager [req-dd80fc31-19d0-4bf1-b5a2-7332d8520429 req-dae18f41-932a-46db-a8ac-ed56dcced44b service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Received event network-vif-deleted-fe976991-ceca-4699-ad13-b7768f1d9563 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1343.562952] env[61911]: INFO nova.compute.manager [req-dd80fc31-19d0-4bf1-b5a2-7332d8520429 req-dae18f41-932a-46db-a8ac-ed56dcced44b service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Neutron deleted interface fe976991-ceca-4699-ad13-b7768f1d9563; detaching it from the instance and deleting it from the info cache [ 1343.563143] env[61911]: DEBUG nova.network.neutron [req-dd80fc31-19d0-4bf1-b5a2-7332d8520429 req-dae18f41-932a-46db-a8ac-ed56dcced44b service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.038463] env[61911]: DEBUG nova.network.neutron [-] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.066518] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da1776d5-af29-49a3-893c-91a8e256fca6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.075707] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7376af70-f0c5-4dd0-a2c6-81a04bb9f1c2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.098350] env[61911]: DEBUG nova.compute.manager [req-dd80fc31-19d0-4bf1-b5a2-7332d8520429 req-dae18f41-932a-46db-a8ac-ed56dcced44b service nova] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Detach interface failed, port_id=fe976991-ceca-4699-ad13-b7768f1d9563, reason: Instance bc36e378-9d5d-49a5-b7a6-8d4ca3303063 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1344.541119] env[61911]: INFO nova.compute.manager [-] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Took 1.44 seconds to deallocate network for instance. [ 1344.737703] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.737871] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1344.737997] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Rebuilding the list of instances to heal {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1345.048365] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.048698] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1345.048988] env[61911]: DEBUG nova.objects.instance [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'resources' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1345.256926] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1345.257175] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1345.257398] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1345.257626] env[61911]: DEBUG nova.objects.instance [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lazy-loading 'info_cache' on Instance uuid bc36e378-9d5d-49a5-b7a6-8d4ca3303063 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1345.581649] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122977bb-d3b5-4427-af8a-1b300fd06afc {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.589036] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febfb41c-a739-475a-92de-e56cd5ddf76d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.617988] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ebfc6c-cb28-4137-b267-9eb5184c7a5c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.624342] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d6420b-ff2e-4b00-90c9-e4b028d410a5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.636896] env[61911]: DEBUG nova.compute.provider_tree [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1346.139830] env[61911]: DEBUG nova.scheduler.client.report [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1346.279817] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1346.644710] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.596s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1346.664308] env[61911]: INFO nova.scheduler.client.report [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Deleted allocations for instance bc36e378-9d5d-49a5-b7a6-8d4ca3303063 [ 1346.851528] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1347.172015] env[61911]: DEBUG oslo_concurrency.lockutils [None req-2733379c-7e66-4d3b-ae2b-8d91cdea1b4d tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "bc36e378-9d5d-49a5-b7a6-8d4ca3303063" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.710s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1347.354412] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-bc36e378-9d5d-49a5-b7a6-8d4ca3303063" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1347.354609] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1347.354854] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1347.355062] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1347.355199] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1349.614047] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1349.614047] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1350.115610] env[61911]: DEBUG nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Starting instance... {{(pid=61911) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1350.349712] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.636339] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1350.636614] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1350.638123] env[61911]: INFO nova.compute.claims [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1351.671785] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f20faac-5453-4a89-994f-34b447ad49bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.679551] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bee479-8e6b-4ef6-b581-f2bbe0386726 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.707982] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86698129-9511-4690-ab8a-d7d7951c091b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.714667] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e1eb7b-ae07-4c2e-90d2-38558ba8e518 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.728109] env[61911]: DEBUG nova.compute.provider_tree [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1352.231670] env[61911]: DEBUG nova.scheduler.client.report [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1352.737265] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.100s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.737821] env[61911]: DEBUG nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Start building networks asynchronously for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1353.243441] env[61911]: DEBUG nova.compute.utils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1353.245299] env[61911]: DEBUG nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Allocating IP information in the background. {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1353.245480] env[61911]: DEBUG nova.network.neutron [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] allocate_for_instance() {{(pid=61911) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1353.304099] env[61911]: DEBUG nova.policy [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a874ae493bb84b848c196eb62c787bdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '104053ff00ae487692e030d91965a714', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61911) authorize /opt/stack/nova/nova/policy.py:201}} [ 1353.588377] env[61911]: DEBUG nova.network.neutron [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Successfully created port: 22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1353.748218] env[61911]: DEBUG nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Start building block device mappings for instance. {{(pid=61911) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1354.758438] env[61911]: DEBUG nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Start spawning the instance on the hypervisor. {{(pid=61911) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1354.782902] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:52:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:52:15Z,direct_url=,disk_format='vmdk',id=0dfe2ff1-43fd-4529-93f1-daaccc0711cf,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='70794eff7a2c478b800918ba4144a52b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:52:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1354.783163] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Flavor limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1354.783332] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Image limits 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1354.783521] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Flavor pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1354.783686] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Image pref 0:0:0 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1354.783863] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61911) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1354.784093] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1354.784267] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1354.784441] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Got 1 possible topologies {{(pid=61911) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1354.784611] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1354.784792] env[61911]: DEBUG nova.virt.hardware [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61911) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1354.785663] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5784fe40-4916-44eb-aa23-1d2e5a4a83ab {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.793407] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7754da44-123d-4ce9-a77f-c182cc0e547f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.950045] env[61911]: DEBUG nova.compute.manager [req-0ff92ebd-88e0-4c4c-a1b8-ad50be300954 req-d0b87f2c-d375-4b40-9436-5b34fd72bd99 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Received event network-vif-plugged-22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1354.950280] env[61911]: DEBUG oslo_concurrency.lockutils [req-0ff92ebd-88e0-4c4c-a1b8-ad50be300954 req-d0b87f2c-d375-4b40-9436-5b34fd72bd99 service nova] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.950517] env[61911]: DEBUG oslo_concurrency.lockutils [req-0ff92ebd-88e0-4c4c-a1b8-ad50be300954 req-d0b87f2c-d375-4b40-9436-5b34fd72bd99 service nova] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.950758] env[61911]: DEBUG oslo_concurrency.lockutils [req-0ff92ebd-88e0-4c4c-a1b8-ad50be300954 req-d0b87f2c-d375-4b40-9436-5b34fd72bd99 service nova] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1354.950943] env[61911]: DEBUG nova.compute.manager [req-0ff92ebd-88e0-4c4c-a1b8-ad50be300954 req-d0b87f2c-d375-4b40-9436-5b34fd72bd99 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] No waiting events found dispatching network-vif-plugged-22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1354.951132] env[61911]: WARNING nova.compute.manager [req-0ff92ebd-88e0-4c4c-a1b8-ad50be300954 req-d0b87f2c-d375-4b40-9436-5b34fd72bd99 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Received unexpected event network-vif-plugged-22431b53-2789-47c6-9a1e-7508b4915147 for instance with vm_state building and task_state spawning. [ 1355.030558] env[61911]: DEBUG nova.network.neutron [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Successfully updated port: 22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1355.534095] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1355.534095] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1355.534095] env[61911]: DEBUG nova.network.neutron [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Building network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1356.072363] env[61911]: DEBUG nova.network.neutron [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Instance cache missing network info. {{(pid=61911) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1356.200938] env[61911]: DEBUG nova.network.neutron [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating instance_info_cache with network_info: [{"id": "22431b53-2789-47c6-9a1e-7508b4915147", "address": "fa:16:3e:22:26:e6", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22431b53-27", "ovs_interfaceid": "22431b53-2789-47c6-9a1e-7508b4915147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1356.703856] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1356.704212] env[61911]: DEBUG nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Instance network_info: |[{"id": "22431b53-2789-47c6-9a1e-7508b4915147", "address": "fa:16:3e:22:26:e6", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22431b53-27", "ovs_interfaceid": "22431b53-2789-47c6-9a1e-7508b4915147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61911) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1356.704678] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:26:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1fb81f98-6f5a-47ab-a512-27277591d064', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22431b53-2789-47c6-9a1e-7508b4915147', 'vif_model': 'vmxnet3'}] {{(pid=61911) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1356.712091] env[61911]: DEBUG oslo.service.loopingcall [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1356.712318] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Creating VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1356.712540] env[61911]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56ba25ad-f4df-47e2-801e-bd58499ca184 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.732155] env[61911]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1356.732155] env[61911]: value = "task-1251795" [ 1356.732155] env[61911]: _type = "Task" [ 1356.732155] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.739304] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251795, 'name': CreateVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.974453] env[61911]: DEBUG nova.compute.manager [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Received event network-changed-22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1356.974635] env[61911]: DEBUG nova.compute.manager [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Refreshing instance network info cache due to event network-changed-22431b53-2789-47c6-9a1e-7508b4915147. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1356.974860] env[61911]: DEBUG oslo_concurrency.lockutils [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] Acquiring lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1356.975014] env[61911]: DEBUG oslo_concurrency.lockutils [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] Acquired lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1356.975193] env[61911]: DEBUG nova.network.neutron [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Refreshing network info cache for port 22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1357.243052] env[61911]: DEBUG oslo_vmware.api [-] Task: {'id': task-1251795, 'name': CreateVM_Task, 'duration_secs': 0.277604} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.243443] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Created VM on the ESX host {{(pid=61911) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1357.243919] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.244106] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.244427] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1357.244681] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3b2ea0c-4e54-4837-8485-44ae4da350bd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.249495] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1357.249495] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d6e144-8c95-17c9-96e5-b31dde18cb8b" [ 1357.249495] env[61911]: _type = "Task" [ 1357.249495] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.256798] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d6e144-8c95-17c9-96e5-b31dde18cb8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.675959] env[61911]: DEBUG nova.network.neutron [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updated VIF entry in instance network info cache for port 22431b53-2789-47c6-9a1e-7508b4915147. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1357.676386] env[61911]: DEBUG nova.network.neutron [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating instance_info_cache with network_info: [{"id": "22431b53-2789-47c6-9a1e-7508b4915147", "address": "fa:16:3e:22:26:e6", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22431b53-27", "ovs_interfaceid": "22431b53-2789-47c6-9a1e-7508b4915147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1357.759477] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52d6e144-8c95-17c9-96e5-b31dde18cb8b, 'name': SearchDatastore_Task, 'duration_secs': 0.009867} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.759729] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1357.759971] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Processing image 0dfe2ff1-43fd-4529-93f1-daaccc0711cf {{(pid=61911) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1357.760229] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.760380] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.760584] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1357.760857] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96f706a0-d735-446a-8429-3a366aa0adc5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.768180] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61911) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1357.768358] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61911) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1357.769015] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b336f43e-23d5-4135-ba6a-52f34d2c57b0 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.773715] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1357.773715] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de3ce4-2e98-f6d8-72a2-0031ea980416" [ 1357.773715] env[61911]: _type = "Task" [ 1357.773715] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.780465] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de3ce4-2e98-f6d8-72a2-0031ea980416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.179257] env[61911]: DEBUG oslo_concurrency.lockutils [req-1fbf0d22-eff2-4e0d-9eb7-3459072449cd req-3423b4ef-7d36-403f-a0df-b6b678c455ab service nova] Releasing lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1358.283731] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52de3ce4-2e98-f6d8-72a2-0031ea980416, 'name': SearchDatastore_Task, 'duration_secs': 0.007578} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.284528] env[61911]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c731586-f702-488e-b43d-4a84b0865ce1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.289322] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1358.289322] env[61911]: value = "session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e05641-e409-6de7-3dbb-f1d6bde6cfce" [ 1358.289322] env[61911]: _type = "Task" [ 1358.289322] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.296676] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e05641-e409-6de7-3dbb-f1d6bde6cfce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.799653] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': session[529e3ea1-81cb-5b13-81cb-4b5b8ffadaea]52e05641-e409-6de7-3dbb-f1d6bde6cfce, 'name': SearchDatastore_Task, 'duration_secs': 0.009233} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.799925] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1358.800208] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] ca40de1b-3e1f-459e-9ae7-bc31eb472828/ca40de1b-3e1f-459e-9ae7-bc31eb472828.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1358.800465] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dac1945e-bdc4-4509-8c7a-908860f9b74b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.806438] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1358.806438] env[61911]: value = "task-1251796" [ 1358.806438] env[61911]: _type = "Task" [ 1358.806438] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.814832] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.315969] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251796, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.41829} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.316787] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0dfe2ff1-43fd-4529-93f1-daaccc0711cf/0dfe2ff1-43fd-4529-93f1-daaccc0711cf.vmdk to [datastore1] ca40de1b-3e1f-459e-9ae7-bc31eb472828/ca40de1b-3e1f-459e-9ae7-bc31eb472828.vmdk {{(pid=61911) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1359.317207] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Extending root virtual disk to 1048576 {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1359.317590] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49382c6b-37c2-4f48-a519-cd28c23e0bbf {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.323527] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1359.323527] env[61911]: value = "task-1251797" [ 1359.323527] env[61911]: _type = "Task" [ 1359.323527] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.330378] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251797, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.833095] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251797, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057402} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.833365] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Extended root virtual disk {{(pid=61911) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1359.834117] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630085dd-3060-4b5e-a220-3ebbe78d7c72 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.854965] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] ca40de1b-3e1f-459e-9ae7-bc31eb472828/ca40de1b-3e1f-459e-9ae7-bc31eb472828.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1359.855207] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d68dc4ab-24d2-4dd0-84b8-80416d550206 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.874101] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1359.874101] env[61911]: value = "task-1251798" [ 1359.874101] env[61911]: _type = "Task" [ 1359.874101] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.881404] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251798, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.383862] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251798, 'name': ReconfigVM_Task, 'duration_secs': 0.283684} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.384217] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfigured VM instance instance-0000006e to attach disk [datastore1] ca40de1b-3e1f-459e-9ae7-bc31eb472828/ca40de1b-3e1f-459e-9ae7-bc31eb472828.vmdk or device None with type sparse {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1360.384800] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffa0eec2-5884-427e-b75d-f42dd001dc01 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.390600] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1360.390600] env[61911]: value = "task-1251799" [ 1360.390600] env[61911]: _type = "Task" [ 1360.390600] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.397566] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251799, 'name': Rename_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.899956] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251799, 'name': Rename_Task, 'duration_secs': 0.1457} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.900258] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Powering on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1360.900504] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9fe58cb-3e95-431d-8419-6c3a0c9c5c4f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.906886] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1360.906886] env[61911]: value = "task-1251800" [ 1360.906886] env[61911]: _type = "Task" [ 1360.906886] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.913911] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251800, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.416557] env[61911]: DEBUG oslo_vmware.api [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251800, 'name': PowerOnVM_Task, 'duration_secs': 0.427598} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.416932] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Powered on the VM {{(pid=61911) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1361.417027] env[61911]: INFO nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Took 6.66 seconds to spawn the instance on the hypervisor. [ 1361.417224] env[61911]: DEBUG nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Checking state {{(pid=61911) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1361.417959] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2a4d54-6172-4732-8889-00b87f50a185 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.935508] env[61911]: INFO nova.compute.manager [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Took 11.31 seconds to build instance. [ 1362.438201] env[61911]: DEBUG oslo_concurrency.lockutils [None req-f1d281ab-3baa-4a83-9cdc-4f08e70d3f67 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.824s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1362.630549] env[61911]: DEBUG nova.compute.manager [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Received event network-changed-22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1362.630846] env[61911]: DEBUG nova.compute.manager [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Refreshing instance network info cache due to event network-changed-22431b53-2789-47c6-9a1e-7508b4915147. {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1362.631127] env[61911]: DEBUG oslo_concurrency.lockutils [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] Acquiring lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1362.631289] env[61911]: DEBUG oslo_concurrency.lockutils [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] Acquired lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1362.631460] env[61911]: DEBUG nova.network.neutron [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Refreshing network info cache for port 22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1363.340921] env[61911]: DEBUG nova.network.neutron [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updated VIF entry in instance network info cache for port 22431b53-2789-47c6-9a1e-7508b4915147. {{(pid=61911) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1363.341329] env[61911]: DEBUG nova.network.neutron [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating instance_info_cache with network_info: [{"id": "22431b53-2789-47c6-9a1e-7508b4915147", "address": "fa:16:3e:22:26:e6", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22431b53-27", "ovs_interfaceid": "22431b53-2789-47c6-9a1e-7508b4915147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1363.844109] env[61911]: DEBUG oslo_concurrency.lockutils [req-330c2306-a6e1-46a2-a347-5468e15fefed req-a2375010-7f62-4d0e-a1ce-aa2ed7022c69 service nova] Releasing lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1398.738377] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1399.241185] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.241414] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1399.241588] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1399.241742] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1399.242688] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077083b6-91ca-4cad-9dd0-4b9d3e8d65f9 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.250712] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240e1380-b400-4035-8c03-40ff35eb1a1b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.263856] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4931aa1e-f08c-41ae-8541-9ee38ccbe338 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.270012] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72e40ad-0cea-4f60-b10e-a00e09e18bfe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.297249] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181464MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1399.297384] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.297573] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1400.234486] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1400.234783] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1400.323488] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Instance ca40de1b-3e1f-459e-9ae7-bc31eb472828 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61911) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1400.323720] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1400.323868] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1400.349206] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c176af-4b3b-42bd-8d7e-0e351a4261dd {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.356977] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871f3f4d-59ae-44e9-aee3-6f5cd442e410 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.386533] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67d8d14-6114-448b-ad94-3a2796409048 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.393436] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5242a0e4-a901-496d-bf4c-5bb818d6df04 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.406036] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1400.737555] env[61911]: DEBUG nova.compute.utils [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1400.909098] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1401.240854] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1401.414961] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61911) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1401.415258] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.118s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1401.415454] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.296563] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1402.296965] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1402.297091] env[61911]: INFO nova.compute.manager [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Attaching volume 7d0df307-9879-4fab-82aa-b6ab7f6c52e7 to /dev/sdb [ 1402.326781] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9ae964-b795-4e29-bc22-55351d06161e {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.334102] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c578f8d7-79cc-4bbf-888f-fd9729bc4aa8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.346897] env[61911]: DEBUG nova.virt.block_device [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating existing volume attachment record: 125d7168-6b78-4aad-bd23-129e04844101 {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1402.911071] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.911071] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.911071] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.911071] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.911350] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.911350] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Cleaning up deleted instances {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11608}} [ 1403.416515] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] There are 7 instances to clean {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11617}} [ 1403.416885] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: bc36e378-9d5d-49a5-b7a6-8d4ca3303063] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1403.920578] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 90c0b35d-430a-4946-8a02-bae7f575e5ef] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1404.423756] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 2077036f-3451-416c-8783-aa3ea5fbcbcb] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1404.927129] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: a3596e80-37b9-4e7b-bf6e-8f995df1deed] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1405.430215] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 8ac6fded-77cd-4258-8fcf-c36f27435c20] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1405.934062] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 1198724e-2802-47a0-a7e8-3941e3ffc175] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1406.437625] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: 95853730-1908-42f4-b30f-a4f276d2da8b] Instance has had 0 of 5 cleanup attempts {{(pid=61911) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1406.889485] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1406.889729] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269706', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'name': 'volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'serial': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1406.890667] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999c2ebf-ff2c-4c9d-a24d-0545a1a135b3 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.906996] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b6846f-ed21-4a60-9511-7031586fba55 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.930629] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7/volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1406.930869] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfab0109-ac68-43b9-a79a-c65d85189977 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.949645] env[61911]: DEBUG oslo_vmware.api [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1406.949645] env[61911]: value = "task-1251803" [ 1406.949645] env[61911]: _type = "Task" [ 1406.949645] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.957301] env[61911]: DEBUG oslo_vmware.api [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251803, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.459355] env[61911]: DEBUG oslo_vmware.api [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251803, 'name': ReconfigVM_Task, 'duration_secs': 0.321385} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.459771] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7/volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1407.464359] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a27a5002-1f73-40bf-869e-718770dfa2f5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.478504] env[61911]: DEBUG oslo_vmware.api [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1407.478504] env[61911]: value = "task-1251804" [ 1407.478504] env[61911]: _type = "Task" [ 1407.478504] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.487450] env[61911]: DEBUG oslo_vmware.api [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251804, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.737815] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.738031] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Starting heal instance info cache {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1407.738137] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Rebuilding the list of instances to heal {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1407.987939] env[61911]: DEBUG oslo_vmware.api [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251804, 'name': ReconfigVM_Task, 'duration_secs': 0.127453} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.988235] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269706', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'name': 'volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'serial': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1408.283602] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1408.283807] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquired lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1408.283979] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Forcefully refreshing network info cache for instance {{(pid=61911) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1408.284154] env[61911]: DEBUG nova.objects.instance [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lazy-loading 'info_cache' on Instance uuid ca40de1b-3e1f-459e-9ae7-bc31eb472828 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1409.023787] env[61911]: DEBUG nova.objects.instance [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid ca40de1b-3e1f-459e-9ae7-bc31eb472828 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1409.530225] env[61911]: DEBUG oslo_concurrency.lockutils [None req-9f2f17b8-0944-40b6-afb3-379eef83c1fd tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.233s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1410.009933] env[61911]: DEBUG nova.network.neutron [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating instance_info_cache with network_info: [{"id": "22431b53-2789-47c6-9a1e-7508b4915147", "address": "fa:16:3e:22:26:e6", "network": {"id": "7d6f1ed8-1e09-4c0a-b017-0666c292a8fb", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1642998733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104053ff00ae487692e030d91965a714", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22431b53-27", "ovs_interfaceid": "22431b53-2789-47c6-9a1e-7508b4915147", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1410.341489] env[61911]: DEBUG oslo_concurrency.lockutils [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1410.341764] env[61911]: DEBUG oslo_concurrency.lockutils [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1410.512294] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Releasing lock "refresh_cache-ca40de1b-3e1f-459e-9ae7-bc31eb472828" {{(pid=61911) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1410.512466] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updated the network info_cache for instance {{(pid=61911) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1410.512645] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1410.512808] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1410.512965] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1410.513106] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61911) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1410.513271] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1410.513392] env[61911]: DEBUG nova.compute.manager [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Cleaning up deleted instances with incomplete migration {{(pid=61911) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11646}} [ 1410.844967] env[61911]: DEBUG nova.compute.utils [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Using /dev/sd instead of None {{(pid=61911) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1411.348342] env[61911]: DEBUG oslo_concurrency.lockutils [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1412.403542] env[61911]: DEBUG oslo_concurrency.lockutils [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1412.403925] env[61911]: DEBUG oslo_concurrency.lockutils [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1412.404122] env[61911]: INFO nova.compute.manager [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Attaching volume f5e6a873-681d-4a60-aebe-07bae17ad69d to /dev/sdc [ 1412.434142] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19d3095-f697-4533-98f4-608c45bc6882 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.441235] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3701ef-91e0-43d0-9fbd-5b90792e3b98 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.454408] env[61911]: DEBUG nova.virt.block_device [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating existing volume attachment record: 15330978-bfc9-4cc9-ac1d-a65402e2468a {{(pid=61911) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1416.996951] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Volume attach. Driver type: vmdk {{(pid=61911) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1416.997281] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269707', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'name': 'volume-f5e6a873-681d-4a60-aebe-07bae17ad69d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'serial': 'f5e6a873-681d-4a60-aebe-07bae17ad69d'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1416.998293] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c50bb1-fe81-45f1-830a-d80c3788e65f {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.016167] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3436ecf3-fd01-4b27-8d13-120a9bf7dde4 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.046608] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-f5e6a873-681d-4a60-aebe-07bae17ad69d/volume-f5e6a873-681d-4a60-aebe-07bae17ad69d.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1417.046874] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ff9939f-21be-4998-8033-9916abfc1342 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.065530] env[61911]: DEBUG oslo_vmware.api [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1417.065530] env[61911]: value = "task-1251807" [ 1417.065530] env[61911]: _type = "Task" [ 1417.065530] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.073262] env[61911]: DEBUG oslo_vmware.api [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251807, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.575802] env[61911]: DEBUG oslo_vmware.api [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251807, 'name': ReconfigVM_Task, 'duration_secs': 0.347587} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.576104] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-f5e6a873-681d-4a60-aebe-07bae17ad69d/volume-f5e6a873-681d-4a60-aebe-07bae17ad69d.vmdk or device None with type thin {{(pid=61911) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1417.580657] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff4dc73e-59b8-4d8a-b871-d5d27fe8c9a2 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.594335] env[61911]: DEBUG oslo_vmware.api [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1417.594335] env[61911]: value = "task-1251808" [ 1417.594335] env[61911]: _type = "Task" [ 1417.594335] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.601506] env[61911]: DEBUG oslo_vmware.api [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251808, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.104128] env[61911]: DEBUG oslo_vmware.api [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251808, 'name': ReconfigVM_Task, 'duration_secs': 0.135394} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.104454] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269707', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'name': 'volume-f5e6a873-681d-4a60-aebe-07bae17ad69d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'serial': 'f5e6a873-681d-4a60-aebe-07bae17ad69d'} {{(pid=61911) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1419.138990] env[61911]: DEBUG nova.objects.instance [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid ca40de1b-3e1f-459e-9ae7-bc31eb472828 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1419.644420] env[61911]: DEBUG oslo_concurrency.lockutils [None req-00050f87-18d6-4f5e-9aaf-ff40b3e90137 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1419.936707] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1419.937097] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1420.440522] env[61911]: INFO nova.compute.manager [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Detaching volume 7d0df307-9879-4fab-82aa-b6ab7f6c52e7 [ 1420.470456] env[61911]: INFO nova.virt.block_device [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Attempting to driver detach volume 7d0df307-9879-4fab-82aa-b6ab7f6c52e7 from mountpoint /dev/sdb [ 1420.470708] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1420.470902] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269706', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'name': 'volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'serial': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1420.471791] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa6e4d9-4fed-46bd-85f6-87af83035ab6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.495971] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc3a3b2-acd4-43b8-b6ca-6405c7a8d54d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.502516] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d8d6e1-bb07-4703-b683-73368c3248d6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.525879] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1fda393-3da4-43db-ae27-6496e0bfaf22 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.539760] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] The volume has not been displaced from its original location: [datastore1] volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7/volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1420.544907] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1420.545220] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1c26884-798f-4f7e-ad28-93f4439f1b1b {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.561709] env[61911]: DEBUG oslo_vmware.api [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1420.561709] env[61911]: value = "task-1251809" [ 1420.561709] env[61911]: _type = "Task" [ 1420.561709] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.568786] env[61911]: DEBUG oslo_vmware.api [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251809, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.071560] env[61911]: DEBUG oslo_vmware.api [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251809, 'name': ReconfigVM_Task, 'duration_secs': 0.221853} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.071849] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1421.076355] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57bbdad4-5cd1-41a2-859f-62111b7d0a32 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.090125] env[61911]: DEBUG oslo_vmware.api [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1421.090125] env[61911]: value = "task-1251810" [ 1421.090125] env[61911]: _type = "Task" [ 1421.090125] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.099157] env[61911]: DEBUG oslo_vmware.api [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251810, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.599336] env[61911]: DEBUG oslo_vmware.api [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251810, 'name': ReconfigVM_Task, 'duration_secs': 0.125206} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.599718] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269706', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'name': 'volume-7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7', 'serial': '7d0df307-9879-4fab-82aa-b6ab7f6c52e7'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1422.139304] env[61911]: DEBUG nova.objects.instance [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid ca40de1b-3e1f-459e-9ae7-bc31eb472828 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1423.146480] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a78c17ae-0049-4143-b357-40002104d930 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.209s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.168296] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.168533] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.671630] env[61911]: INFO nova.compute.manager [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Detaching volume f5e6a873-681d-4a60-aebe-07bae17ad69d [ 1423.700588] env[61911]: INFO nova.virt.block_device [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Attempting to driver detach volume f5e6a873-681d-4a60-aebe-07bae17ad69d from mountpoint /dev/sdc [ 1423.700837] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Volume detach. Driver type: vmdk {{(pid=61911) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1423.701040] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269707', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'name': 'volume-f5e6a873-681d-4a60-aebe-07bae17ad69d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'serial': 'f5e6a873-681d-4a60-aebe-07bae17ad69d'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1423.702321] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd47241-f912-4736-852f-430fb41205c6 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.722808] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10300ac2-23d6-4e57-98d8-4fa03ec8a5ba {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.729333] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1d8e9e-7b5e-4ee0-9dc0-8b8bcba1128d {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.748269] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22672ad5-9092-4832-97a3-f47c4b1bca13 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.761805] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] The volume has not been displaced from its original location: [datastore1] volume-f5e6a873-681d-4a60-aebe-07bae17ad69d/volume-f5e6a873-681d-4a60-aebe-07bae17ad69d.vmdk. No consolidation needed. {{(pid=61911) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1423.766899] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfiguring VM instance instance-0000006e to detach disk 2002 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1423.767161] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8465925-55bf-429f-9ad2-9d6cf316db55 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.783409] env[61911]: DEBUG oslo_vmware.api [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1423.783409] env[61911]: value = "task-1251811" [ 1423.783409] env[61911]: _type = "Task" [ 1423.783409] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.790456] env[61911]: DEBUG oslo_vmware.api [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251811, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.293541] env[61911]: DEBUG oslo_vmware.api [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251811, 'name': ReconfigVM_Task, 'duration_secs': 0.203858} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.293919] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Reconfigured VM instance instance-0000006e to detach disk 2002 {{(pid=61911) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1424.298399] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c07d6636-a80f-4394-8613-6e1b90f78f54 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.315609] env[61911]: DEBUG oslo_vmware.api [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1424.315609] env[61911]: value = "task-1251812" [ 1424.315609] env[61911]: _type = "Task" [ 1424.315609] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.323912] env[61911]: DEBUG oslo_vmware.api [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251812, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.825264] env[61911]: DEBUG oslo_vmware.api [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251812, 'name': ReconfigVM_Task, 'duration_secs': 0.12498} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.825566] env[61911]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269707', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'name': 'volume-f5e6a873-681d-4a60-aebe-07bae17ad69d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca40de1b-3e1f-459e-9ae7-bc31eb472828', 'attached_at': '', 'detached_at': '', 'volume_id': 'f5e6a873-681d-4a60-aebe-07bae17ad69d', 'serial': 'f5e6a873-681d-4a60-aebe-07bae17ad69d'} {{(pid=61911) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1425.366274] env[61911]: DEBUG nova.objects.instance [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'flavor' on Instance uuid ca40de1b-3e1f-459e-9ae7-bc31eb472828 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1426.374515] env[61911]: DEBUG oslo_concurrency.lockutils [None req-3b7a3398-e186-44e8-abfc-090bb78363bb tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.206s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1427.523655] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1427.524175] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1427.524434] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1427.524724] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1427.524988] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1427.527265] env[61911]: INFO nova.compute.manager [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Terminating instance [ 1428.030972] env[61911]: DEBUG nova.compute.manager [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Start destroying the instance on the hypervisor. {{(pid=61911) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1428.031247] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Destroying instance {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1428.032179] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8364a0eb-6ae8-4dd1-bb2b-32b677190cf8 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.039872] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Powering off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1428.040107] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07472f7a-cac7-44c8-b459-9876939fbc97 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.046161] env[61911]: DEBUG oslo_vmware.api [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1428.046161] env[61911]: value = "task-1251813" [ 1428.046161] env[61911]: _type = "Task" [ 1428.046161] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.053996] env[61911]: DEBUG oslo_vmware.api [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251813, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1428.556810] env[61911]: DEBUG oslo_vmware.api [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251813, 'name': PowerOffVM_Task, 'duration_secs': 0.191476} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1428.557209] env[61911]: DEBUG nova.virt.vmwareapi.vm_util [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Powered off the VM {{(pid=61911) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1428.557330] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Unregistering the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1428.557513] env[61911]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cab3fa7d-cb5b-41a5-b4e2-d23bfbb6a199 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.616672] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Unregistered the VM {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1428.616916] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Deleting contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1428.617063] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Deleting the datastore file [datastore1] ca40de1b-3e1f-459e-9ae7-bc31eb472828 {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1428.617329] env[61911]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-438fe2e6-453e-4d3a-820c-5b9616119a85 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.623362] env[61911]: DEBUG oslo_vmware.api [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for the task: (returnval){ [ 1428.623362] env[61911]: value = "task-1251815" [ 1428.623362] env[61911]: _type = "Task" [ 1428.623362] env[61911]: } to complete. {{(pid=61911) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1428.630713] env[61911]: DEBUG oslo_vmware.api [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1429.133740] env[61911]: DEBUG oslo_vmware.api [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Task: {'id': task-1251815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127233} completed successfully. {{(pid=61911) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1429.134115] env[61911]: DEBUG nova.virt.vmwareapi.ds_util [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Deleted the datastore file {{(pid=61911) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1429.134337] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Deleted contents of the VM from datastore datastore1 {{(pid=61911) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1429.134532] env[61911]: DEBUG nova.virt.vmwareapi.vmops [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Instance destroyed {{(pid=61911) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1429.134720] env[61911]: INFO nova.compute.manager [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1429.134971] env[61911]: DEBUG oslo.service.loopingcall [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61911) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1429.135189] env[61911]: DEBUG nova.compute.manager [-] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Deallocating network for instance {{(pid=61911) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1429.135289] env[61911]: DEBUG nova.network.neutron [-] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] deallocate_for_instance() {{(pid=61911) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1429.610532] env[61911]: DEBUG nova.compute.manager [req-3e2ee014-cbc2-4478-a8b6-ca471926d4b6 req-5802540d-d32b-4ca5-85f6-0419135feaf8 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Received event network-vif-deleted-22431b53-2789-47c6-9a1e-7508b4915147 {{(pid=61911) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1429.611158] env[61911]: INFO nova.compute.manager [req-3e2ee014-cbc2-4478-a8b6-ca471926d4b6 req-5802540d-d32b-4ca5-85f6-0419135feaf8 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Neutron deleted interface 22431b53-2789-47c6-9a1e-7508b4915147; detaching it from the instance and deleting it from the info cache [ 1429.611369] env[61911]: DEBUG nova.network.neutron [req-3e2ee014-cbc2-4478-a8b6-ca471926d4b6 req-5802540d-d32b-4ca5-85f6-0419135feaf8 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1430.086868] env[61911]: DEBUG nova.network.neutron [-] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Updating instance_info_cache with network_info: [] {{(pid=61911) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1430.113624] env[61911]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad67e71d-4fdc-4c02-8b2c-25d790e01135 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.124885] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329e72d0-5fa1-4497-9d1b-be1c2d2a2ad5 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.147146] env[61911]: DEBUG nova.compute.manager [req-3e2ee014-cbc2-4478-a8b6-ca471926d4b6 req-5802540d-d32b-4ca5-85f6-0419135feaf8 service nova] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Detach interface failed, port_id=22431b53-2789-47c6-9a1e-7508b4915147, reason: Instance ca40de1b-3e1f-459e-9ae7-bc31eb472828 could not be found. {{(pid=61911) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1430.590139] env[61911]: INFO nova.compute.manager [-] [instance: ca40de1b-3e1f-459e-9ae7-bc31eb472828] Took 1.45 seconds to deallocate network for instance. [ 1431.096920] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1431.097319] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1431.097439] env[61911]: DEBUG nova.objects.instance [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lazy-loading 'resources' on Instance uuid ca40de1b-3e1f-459e-9ae7-bc31eb472828 {{(pid=61911) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1431.722934] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491ba2cf-ae59-4bfd-b0d1-7129ae3f9504 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.730530] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09afce99-8202-4fed-b8a3-5c3bfabc832a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.759396] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34546bc7-1eae-4976-9440-ce9d57ce3719 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.765752] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361759fc-e26b-4149-94e7-fbc61a6399e1 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.778120] env[61911]: DEBUG nova.compute.provider_tree [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1432.282108] env[61911]: DEBUG nova.scheduler.client.report [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1432.786960] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1432.809140] env[61911]: INFO nova.scheduler.client.report [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Deleted allocations for instance ca40de1b-3e1f-459e-9ae7-bc31eb472828 [ 1433.317348] env[61911]: DEBUG oslo_concurrency.lockutils [None req-319c47fe-142d-48b9-b2b4-e30e0451a760 tempest-AttachVolumeTestJSON-391738973 tempest-AttachVolumeTestJSON-391738973-project-member] Lock "ca40de1b-3e1f-459e-9ae7-bc31eb472828" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.793s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1461.240072] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1461.240486] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1461.240486] env[61911]: DEBUG oslo_service.periodic_task [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61911) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1461.743560] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1461.743799] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1461.743965] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1461.744155] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61911) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1461.745124] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd11ee5-29e4-4690-84fc-a5bfcc3ef0fe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.753487] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46d1caa-eef3-44c1-9de3-eb15a6e23990 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.767044] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f740da4-f22e-45e2-97ec-115c1e6b1826 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.772698] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6dd3434-5361-42e4-bdb0-856d7c1f689c {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1461.801595] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181379MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61911) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1461.801740] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1461.801925] env[61911]: DEBUG oslo_concurrency.lockutils [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61911) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1462.821330] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1462.821560] env[61911]: DEBUG nova.compute.resource_tracker [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61911) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1462.835225] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43b2b30-4db9-48ae-8836-557a5daf1afe {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.842616] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed402fc7-c54a-4a47-90be-a272b30109af {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.871132] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc7b90c-50c6-4642-aeac-c8b563ee1039 {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.877821] env[61911]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37527cbc-c8d0-4ba8-924b-6a6e0380ea6a {{(pid=61911) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1462.892277] env[61911]: DEBUG nova.compute.provider_tree [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed in ProviderTree for provider: b8a56394-897a-4e67-92ba-db46db6115e9 {{(pid=61911) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1463.395169] env[61911]: DEBUG nova.scheduler.client.report [None req-a0a81bb9-d74b-4a55-89b1-e12c209d2cf2 None None] Inventory has not changed for provider b8a56394-897a-4e67-92ba-db46db6115e9 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61911) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}}